var/home/core/zuul-output/0000755000175000017500000000000015145722206014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015145725241015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000212741115145725166020272 0ustar corecorevikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9Gf8^HtufX]]6f;l6? ow )S??xI[mEy},fۮWe~7Nû/wb~1;ZxsY~ݳ( 2[$7۫j{Zw鶾z?&~|XLXlN_/:oXx$%X"LADA@@tkޕf{5Wbx=@^J})K3x~JkwI|YowS˷j̶֛]/8 N Rm(of`\r\L>{Jm 0{vR̍>dQQ.aLk~g\UlxDJfw6xi1U2 c#FD?2SgafO3|,ejoLR3[ D HJP1Ub2 e˛X"oWy9j3undd| 0ʡPBU[fi;dYu'Y{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}Pgh+NqɌDSd1d9nT#Abn q1J# !8,$RNI?]BŊ6EZ|^%L[EC 7gg/碓 yi\[.!=A(%Ud,QwC!b]}F*VYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4 wLdW3RG؍:-~<*KmrI,״+i̸.ᑇy ^t }|#qg9b2oII"9 1"6DS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeIVFδ֯`PÝr JovJw|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n@^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.bqMWkB.yhi-cSDCR6"KaFٗt<>vRڡc0SAA\cH{Ⱦ79`®3uO0T-Oy+tǭQI%Q$SiJ. 9F[L1c!zG|k{kEu+Q & ">[@Kfځ9g ? j럚Sř>]uw`C}-{C):fUrG C;Fg@(r8I_|AcH&Y``:"s ayiBq)u%'4 yܽ yW0 -i̭uJ{KưЖ@+UBj -&JO x@}DS.€>3T0|9ē7$3z^.I< )9qf e%dhy:O40n'c}c1XҸuFiƠIkaIx( +")OtZ l^Z^CQ6tffEmDφǽ{QiOENG{P;sHz"G- >+`قSᔙD'Ad [kP+<, {Z5׷!)'xmv0;|!B`0p1y6 PM3rr1TZ')*R ,j>Eeք Dj7NI0[EΰPaySwX7?T4}, l s}ґ4nev$\ ʀu*,L(EĈhz[}&+pzk0%X8Uae 4݃bɩMf8t݃&(sj 44찲XA E#EPV_WWUbU5ì"M|܊W7|}N;6od NN&DǭZrb5Iffe6Rh&C4>Qwf8*c4˥ĘP0W YW ].P!_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGAtĘud7"/6sF%%´ƭ*( :xB_2YK]7.w47mnjGgG{9_e552s4IGx A7yTJ$KOL-aP+;;%+_6'Sr|@2nQ{aK|bjܒ^o(מO80$QxBcXE ءp.֣?5t罦-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAp)vDa;|/֦ I<)tKl3GIĨmIEQ«` RPZ(D2vh>3fsGiZSس:$3w}o$ <S1wg y &SL9qk;NP> ,wդ{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kޏ=S-ܑ2ƹڞ7կZa{0dqw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%<~ X򴱞_aM:E.Qg1DllЊE҉L ehJx{z{tKmdߟ9 &2vA.:Mw~^`X\u6|6rcIF3o!C>Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0aT&O_R-%[ R'l}ʰ (T$ n#b@hpj:˾ojs)M/8`$:) X+ҧSaۥzw}^P1J%+P:Dsƫ%z; +g 0հc0E) 3jƯ?e|-4d {ܶR}>\})'QD([ӠLHٻ J ؖ%J_>7$VO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))OrT 9KaͶ{`3W6I\NYf_Mn6s!9^BG`#hLҡwo˹v>7 _:+$߇v{vzM$VbήdsOw<}#b[E7imH'Y`;5{$ь'gISzp; AI+*ބ(i" 78NBx;c4<ニ)H .Pd^cR^p_G+E--ۥ_F]a|v@|3p%kzh|k*BBRib\J3Yn|뇱[FfP%M:<`pz?]6laz5`ZQs{>3ư_o%oU׆]YLz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd=d@eU XyX2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^u I66|*f\#ߍpo8sx[o%~wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFPȟW֙`mn Lv%7mSU@n_Vۀl9BIcSxlT![`[clzFض˪.l >7l@ΖLl gEj gWUDnr7AG;lU6ieabp៚U|,}S@t1:X  .xI_7ve Z@7IxQpɡr~]Si!ڣZmʢ鉗phw j8\c4>0` R?da,ȍ/ءfQ 2ؐfc}l 2窾ۉ1kw |VvlK۴ymkiK_oK`8 )v3v?נ:b(v6& `-K;~:|F6vXpw*t]r@ 5 ƻ7۱ַ P񷍋 3)Cl^]U҅yY9 &K<-na'Xk,P4+`Þ__e^0)wFO.= w ?>ȑ3n?z,t s5Z/ Clo-` z?a~b mzkC zF/}b&x Uhm.O 4m6^^osVЦ+*@5Fˢg'!>$]0 1_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\o?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nKv8M  /`-u+[w 5쭅]7T=1:T k`8"jѡ6֙ɖ@/#ut{xH9~!24 6e/!~=f)Q UbshY5mseڠ5_mTDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d c(imGMUF>su0,gy(&TI޽*}w[ #j*ٚ- DIAmPvIɦ7聀t>G;_H;2ʗ6 h6QװxmR JQUbTP2j˔Ni)C)HKE"$ӝ!@2<Bq 2oh80,kNA7,?ע|tC3.㤣TiHEIǢƅaeGF$ u2`d)/-st{D1kl)Fa/TTmCԤ{"9b{ywSXE*m#3U ùRIvޏrJ`k|wJKH:O*OKy`( ݢe*{ ua Ȼݔhvׄӫ A^%f+[`sb˟ _.6KsjA Qsmd  O#F.Uf28ZAgy>y,d$C?v01q5e.Um>]RLa&r?+@6k&#l)I5_> ` D s5npo}/ؙq #a2V?X~.4O/'|/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQ^xZb [_tܡ&yЋ{ Sym^?̑sU~' Ԓ f\itu)b>5X -$s?{WƱPz;| \;_D[T/BI GH8@"t*"9E[/Y5d{zrBܖ6Hlc "mKv~[uLU4lZ;xEN'oI㤛rP*jC# 6@dmHg1$ʇȠh#CBΤ{sTQ{%w)7@y1K^ ].Y$46[B-3%OONw8d`Q4d$x0t8@t]y1T\YAidtxBG:pɨyeNg4n]M؞ e}Wn6׳i~'ہZ*FU{fXڃP'Hl4 ,ŸqMHDCYZz Qnz܁$Jp04ȴIL΃.0FiO-qy)i_TA|S2G4miBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ63" ۩:6=TZõ$E,ϓRV|G&$rr;J TtIHFE=RȬ]P pLm|?$%>Eü%mWO[>Xmw,*9.[G n >X8Ī;xW%dT:`ٓ~:QO,}j6j!yڦʲT:Pqҋh] H+&=>g| Z;D8ܶb:! Å{2:+au 6:!fF+0#+̬NY"!6a7#񕪰%:r|o5Znڧs?si/W qEU馥˟^_޶oڷOj'?nc]Rn\t3^邳塨Lɏ"8k8M~?M}OAH$77f|lgn I;.K*!<+"eK5c&`X:#;@B@[(K44sBFu M.MNWLlY]K᜴=/ VމYlϿ4i36$>m|_>9|dUA"{!$jKx E$K3hN(tÊ-#v#O N, 9g80Ǭ&VdӞ5W1!1KYd`,-*&>F~⯰&jb.~cNk BL_OG]Bv.A|'qT(Ol.' 4IE|@Iі)<-p JkQm1 `qacܗVc?)cl*&<}P媠E{-sVU>߇GUt\+n3X]Byoz)li$2cPs6D>TE-n# rve{椱I |p)U݋7yJw&PzDgi xs  xh\L r Ѥo Zt(I >|$>tnMdэoV#ہll/ؽnA(ȱbAj>C9O n6HNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2w⠪R/r| w,?VMqܙ7;qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z]gQ)Bی:D`W&jDk\7XD&?Y\9ȢG:${1`+i n8=%Ml%İȖb7AޗuV3A7ำqE*\qb'YpuHƩҬV nm=Ɂ-2=|5ʹ zi ' ׹U>8bK0%V\ t!Lku`+]c0h&)IVC)p| QUA:]XL/2La[Xѓ F;/-rtx-rei0hE˝ݸDt#{I} `v;jUvK S x1Q2XU&6k&lE"} Q\E)+u>.,SzbQ!g:l0r5aI`"Ǒm O\B!,ZDbjKM%q%Em(>Hm 2z=Eh^&hBk X%t>g:Y #)#vǷOV't d1 =_SEp+%L1OUaY쎹aZNnDZ6fV{r&ȑ|X!|i*FJT+gj׾,$'qg%HWc\4@'@—>9V*E :lw)e6;KK{s`>3X: P/%d1ؑHͦ4;W\hx锎vgqcU!}xF^jc5?7Ua,X nʬ^Cv'A$ƝKA`d;_/EZ~'*"ȜH*Duƽ˳bKg^raͭ̍*tPu*9bJ_ ;3It+v;3O'CX}k:U{⧘pvzz0V Y3'Dco\:^dnJF7a)AH v_§gbȩ<+S%EasUNfB7™:%GY \LXg3۾4\.?}f kj· dM[CaVۿ$XD'QǛU>UݸoRR?x^TE.1߬VwխmLaF݄",Uy%ífz,/o/Z^]ݖF\\UR7򱺹...m/~q[ /7n!7xB[)9nI [GۿsH\ow!>66}եl?|i [%۾s& Z&el-ɬeb.E)բA l1O,dE>-KjLOgeΏe|Bf".ax)֒t0E)J\8ʁ,Gulʂ+lh)6tqd!eó5d ¢ku|M"kP-&ђ5h ^pN0[|B>+q"/[ڲ&6!%<@fpѻKQ31pxFP>TU?!$VQ`Rc1wM "U8V15> =҆#xɮ}U`۸ہt=|X!~Pu(UeS@%Nb:.SZ1d!~\<}LY aBRJ@ѥuȑz.# 3tl7 ]وb Xnݔ[TN1|ttc‡-5=VrPhE0Ǐ}Wd|\aD;(;Ha.]1-{s1`HbKV$n}Z+sz'ʀ*E%N3o2c06JZW?V g>ed\)g.C]pj|4逜*@ nBID f"!!*7kS4޷V+8弔*A19`RI/Hй qPq3TY'퀜+/Ĥ'cp2\1: 0mtH,.7>\hSؗ΀ѩ آSNEYdEcaLF&"FhQ|![gIK v~,Jc%+8[dI368fp*CDrc3k.2WM:UbX[cO;R`RA]d+w!e rr솜[/V`+@;Τ`5d0ϕ_Lع`C"cK>JG.}Ε00e>& 2䯫vNj31c$ i '2Sn-51Y}rE~b>|Ď6Oj~ebIapul9| 3QtUqSCxTD7U9/nq.JYCtuc nrCtVDƖϧ;INOKx%'t+sFUJq:ǫf!NRT1D(3.8Q;І?O+JL0SU%jfˬ1lމZ|VA/.ȍȱh M-r ~[0AG꠭y*8D*-Rz_z{/S[*"꫒?`a;N6uilLn<Yllmb rY״͆jqTI!j.Pٱh s!:W_´KxA|Hk1nE6=W|$O -{]1Ak$ ѫQ6Plp;3F$RveL l5`:~@c>q,7}VE-Q8W70up˳ A¦g/OEU:غA>?=CۣPqȅlW11/$f*0@б 2Dݘrt +qrx!8 J&[V =͋A,z`S,J|L/vrʑ=}IhM4fG(Ȋ1{TT%41Oa'$ 4cj{ƕ_1~lQ)@&4ml)[7z^%ٱ$Rr,|қ.8 oLtFY1Y\S/a"xُ4 o} s7c@88/2,o<%:%'n//'{i|9km쥯&;0mi,A@';ʤDZI^W l)|a07PsLf /Ѥ50-ŸaY?{PhY3GgnAi;|0$_AO%b|tʅ8?#*@,GuO0{Ox 5Lz/H*\2Mװ^gn S ^~~SHC2ڒ$=}OI(/I_BRNo-,"Fu,74q]}ϼGpl4]kZoRh>o/ 9sWk? zWu+ݽ2ԉcf/<*G~^ZP0EU Co0mf/!xD/coGvLݰM_"J$T込F1KcUb }|,?f+Kri[]x)2Bdnq,'6Y<ԟͣZ7yR,x"vX“?ϞH++dc^3a3bG}>iLk&GO)AdepJ "|dA#Z&kHrV'Qh};X[+mÊ+MjSa2{|>".IV$B$](H>.JuKQ~O+]ܿ yv=qMI(h6דIK94v(T7r6g)P@7y΅,+cgIeqR[& s2<>[sPn 'FMӲU<'L0@=J*2=nFAq֦|@C(7UTor.Vm\go(?}){ C=d]޿=K෩5j~@77Qq c*Xyx z ,z)u*1kA{,0 e&%(`"B q,v,B ui8^et!yT.Nҿ9`LRKvOp>u_8G71*v_HcbJr.ÿ/;5S^wt"&ClZpm}<;=Aa@l`5} aO9W:Q`腼+8=Oh<3FIy"w8Gt8w5Qngs< .w.P*LxV,RTo&?5%ֺD&`檽6a1rj-X0}:|DO?vo/l6LCoEmܙu\΄ f\~@?"PFP}.S_[ i|DqgSi iWgX#E8PG> &1&DG*RgM3.:YT%|pEa@dު'gib-qR E .V_RUM.ߝ\ y*Iq^ZA!q>&LjFpMl=TCG#*Uɓۄ2 ߮~SiDȔAH__ô0佺[Zq ,s "˳&\\ U# V7& ,el[rqVmBU} FE^e&0lr&!lfLbk ytR'Fmm//+'ME1mo! 1߼{=4.ʧGj 5آ7RhUR@~hzn&8 M=<>R"m ,7#WW(HYogqD* Ԃ(kh7[Q)@kGF8$ӍΗv90n>3nP]]]ןEAp Mcb7S E}h?0VY3SWAپ_H.>ávBSwazmuhӎUHw|| 5a;hp|#e"Uh[i1q:]$Lh:K*ez/ZY[ R#jGR kv'jv5',lꅠ:8Jfe^L\,"z7\ׯ\q HMmOѵ-Uy})9홟'57轁h eqVQ/p-MYye_p-BgYE:6Auc{AMHvTNHKK:wɼvmK8q]{I&Vy%T U9l(mUx W@?=$~6~dC:Hx֗ZuQY=Gl9jHȥr*/`}Δw$4,=2UT#Ϣ|5(Ӏl7x@K=)GLԞpo%U7눇1k/Le\jeΓ"lLYv`y!voqY۰Ԁ0-oc6ӭ~m[i-'i.7R0uO'qv;ǀQZ!X>iJzMIS& T1PQs%SHsauYHB^K).T:8XseюD9m aW塘#WuQ}Ѽ0w5Hzݲ̆ߴ[wMߗMX)Gj `l95{џ"/rvFt8D#임f.Ωu#*ڔXbre+Bܬz¡"RU]lI\e` ̖n/S2jjdH c]b7in*0iMԚ"*gssy\iaFF;1\PIh`ZUCث. Xl #xnx)X C)~UCP.D^_i$z7l-S/_=c1Wǚϧ3ꓴ qS ]NQ6MFP%Ϡ[er۵fܮJZZ᫊u]dsnWG+ukhZ rյbb ^&\b06@GL5Έi{62]bwM PrmA8ZC?ܮ7oOE6?.ByZJ\'ڋI7&"3@T7:H}NMT7;BuuHKhG$NqH6D:KRTQs}O=nxo&]sKz*פlqNbiV!3}5=\]޼~_ D=~yWy.[*OO=]j .rO~tێÓ:*I_BUdg0ial76C ;S_U ѕ-qv,u؃LCKpVE;jt%nו1:g|{w|QVw3z׎0</TBUG~B5,-n#}P=ͮ:TS Q9c8>D3iB2tX%NMXOKͮ׍f ɪcCW2[XwUcMagaOͮG eT4ey&'ZQg{=KMv"C] R_4KtEHZ]"_rIKwդg(XW16Z?r+58UuV[sm=۸+35v34>߽P|9Uo&ʜDHBjq=)fǷDm&j#~cԦ!6#T |ڶs ha8\{Vj@SѹpΨ!䇏oDLJ92:9SCVb ~RݹCPiN}j9t5dcy*b8xE4?Y1`qn,ݹv(ekڇ00.btөD>ܐNMȥ%Q(=80.Ŕ7y1t|#& Y`T-r3lɀޞ(@_[&g:bS=vgP' tWݜT;1a52Te((2F.Av/#<+fu>M2⒃Ō2OYVBY"ކj̄9fxOcr*cS^w\r0f43,rXAE8 p3p& 9A6>m[+{xI C[ME Krն$;i^CR68*mX_><&-<[w\뿎I2}.6v9(yW?'¦*%_OhG1(׺(MfO:=äf~&)dƦ__T~G~NfB0u7,m?/TV˾Mp,/34owm[)x!PWZ7ϧO(DQ8ED>G6N6?wY&nqGZj#ۓK׻W?(ߍ5Rj7;@,ۖdA'&tYڷ! b7Iآtƀ[u  )ӪMXm:!vE [ l}ɷ=#XSמ(\" T! HLfNDb1 /E\&Y %b"8K <;Og'w㽂XR럊`7!;`f N & EyyJߟ/m@,F@mc8b a%m"(DQ%G# 6#%A3ZdL>eD;ŧD asB,~a%A{P3Da?*p\މJ%'eDdqSfFB()B Cm݈7@,6ƦԪcu #"_KAڍrzFcOw2 w#8R.(:)Qzxk9;]!~P[#F])`F41#GYFw%\=xhg5fҶ cF+6Y\g=)cX[`6](cOu<;+cA!ca /"3=&(P%-GXg0΃>Ф[taCWy@v$qXۍ -TCtpk##!̵p1&b%J.s-vk\[gsa}FGk=S&sJ 3sѡfZ3ZgŭT{kyp vcab=px[ԃb3PLX/>Gb&|KLIrvF7u7 ]/83"B۾&V&Y|[g)20OL<39gpw(>]~g/m&JhWC#90R*5)/B3m˧ ؙ==@}\e'͢6+'!*V+I=Gp7dZ09N;+fbZWҧ 2\[%;2KrBMR pSf^*9KY>qp]wkj( 0;+ONMrCn+Wa&&ʩV jìdE/ n ٶĉOQ^d6x}Æ,x/Z"={\ _!ߝXI6ǪQPsS³PXնBȟ "tQ fu~s&([7~}Wq՚̉SrHndEgamwA`${X} .Z*-3ID$&JxS\Ksѕ-ȩy>{k*5/AKG%igIKM^0~ν d&Kj=*`@]Q{a2Cˆl]x0y 1lYhEyIkp+דY r%nB?*dRL'Ш<'QMp>1, Xo 5F nmɿfz3kE4|fz}xIE x t{M^m4*vr76p]Xv?7܇ln*&H} 'wy֎ٌEzlmѿh!ÔOiVV/Kf19AC%] X?G|) ~73 "'$g+e캾'е Kx'#ytT_;=7*f/ }7[t>&S~w_FEKޚqR S!F 6|D%AR q胞'^ϋ8H_ao=ʤf^Ox4{4Q{y'!h9 DQΖXej|?# "}t]q;I<0h}iѵ`n :ݎ Ҁn X&04$w{]y@}ztW(.wt#V3`A!8 z䦺kZ]`);zCY0I w79BiF7FY23wS1HV(Q@::r 6$&ܰ_pgTQxc](!(拘yGS nb`xbUw^{^P5^5uYר Nc7,"k9YGÿͷD<1kϯnr}h̓=8LʹdzC?4'pʛ=H@aA>w嬄̶l틭Vj֝pRxꍬ Zh!#78f2h~ꈺuSt9OS @XBEi6Iq U<$ O({쑄}RϘG *{⨲+k\7v}>ul=._A,ߟV4R#){A?uIbPbB$TANzO#{${O4BGAhNhO#4x$넆{Oh4BGAhNhFO#4z$BWtcД?O)NPLw|gF{wyWj\ gՃMQNbFnuSSUH12"4e^)4}-S6x݂c>}<{8/ t'4M ޟ9y̪ɯtEe^~$=~QAĠ<>?]>9IG=c|b5Q2շj1/*Вw !YU,T/]@C[ڇE?y?ϫD FVٴ| S-Gr!Ǐe7Sz"/t~hٰ-h[H9#OdCa<_QN&W9߉p$֯_400A+Opˬ1WGA0ΛQ/E*qmy]Nd${p fS ڧS9ߎz+z.=ryDqi>2&iO _UHbE q-I?J~ kC^RTpZ0N]W,0:e0Xd^x&:"2? _o8:I1nS^/*îS dBN؅$a0!֬z 6vGc-H?vT A#z\;iLFfG;&?s/,J8 \9@oK1Gt:{ #C#mTیQ;>BnT3`)*P{m2n<0bͻ6 tNӒwL8@r܎ڗ`?PfJ 5C31޾5) |H)Zvj@GpQ*mH|fTB[6N΍9r{Oxry*M³XW68mQ4ke@t 3T3-Y奰9C;jN%GbZ|e[ViF?s4g 9+ t͜UÃ+8_mKX˂,3BV<򌤅J=C *Lʳ+e(GWprg0ֳTa wcLF蹭S4Vշ{WpT%d!4nqsc7OW(d$&=MIKkAg7`-wzG>/96ˎ#Fm/ Iܘ$$z(1a亘'>õZ1=//F릈Uf>UG5]j#97EM-R/R94ZOϷU}Ll=4 ]DaB~Z] maSs]\޶#ncUbkK>pX69WLRoNcԒ86-fQ[sSGbƯR4&ҴDf]my. m͌nL䶍InOXls,X:%e9sq,6L* J26F~ h;7LEJ;ãJAkIRf:I`ћDLj,`Ⱥ2wJwJVOv;WgCҔ <[*u&IW ER(nnK]p\\5E7FF10LVm2AFI |jl|(2 CZEĠh*,h>" ‡4Ѷ(}4zƍ=|<Ϋ [SPEC#YoP PXHBM~Zvю#ArtŘ)tAr1&6)xIÑ" bkAeM)WLBU@+ܣIڶC$C溻GviL!aZOIIEC0K/BaSHt5Ge[EHRn"1@Iބ"<rsb%D X꧝i47.($Otr@ A<qq|(pӎ1P,kP*aȬ8( @jMn4gu p8tkMc h~ x漚`3eQHx&><>=@΍j{mot稒=-*ْ oGY#OEmClbYFWv}f\M2ؾ?ӆ?Mxc{N1a@^rQ` ajG+n6Oyc_nR˪w]c4ſ}vvGR.o w?߆5p?w'TpU^~–wIʻ ?~wp9w? '5K1OXkb^]w |;~1m6,/$;vײZ;|.m:' bgߗσ@}2me$DsD[Z[ *FdK~H:1џ gH _xwGžb'm'm<10 C0ɏ\Aw֟n:j!Ht Oç.)f큣e{eL29 *P:dxEҩ|~q\=Gw1?9\!ҷߕe[9gm\^@O-[9EL8jxC,jG.j1v['Y|.5Wf75fGLp"wrb@?~ă:#N8)7==g0Zz+)[q%/5'%Htzr٥6u*Ge_yWyI*QZ:`nW y1$"&e vnp,k YЩni`ʟU+S'O ]̑k &Bsc3Vr "jZ 1jtL)Qsj(b`Gt8u[ל>P[; ) YVW͘bH0ULd3 ba&sJ T7$5KۢH:M֪>tQi8.]X#Z&cH->A8t1LkHrù%ޅ /nTU6rDw[sBḬ̂hRehTarG5uGO}z 'O^< Ns*}=cS,YZ6 b66sV)N7.8;zDBVvM6sv B{7C倒 {>ޒ Q&Byߴvֲ+KY }%Ҕ}S䃗GM`)yԧ+%ڹ {.>J7'sG#],jlxkcCmY,Δ2&_ӵA pVsѫʹ"F-׏]pcGlfIwzM$K(coQw_/`^gҤuՓ^>*ROJb UyH)Ϟ,YP6.8~ViG+gIiW6Ge7]Y vl?y\6_oz[je ̫>%cP߿MUͥ]J\!} o.rE>[NK^&TEIL%+&h;-:/*k'/1,oz״8/[ p]Z}zn 3eߢFic\zpQ=7eh3} eJIUr3u]дԣZl;ma=j|r/\epzӺ9xRYB+*c0C[K-8 5xVq"9.8\ͻX; JedNO.eFKXpI{=낟gvRyYmP-֧.+lCj!pӡ۴%Q!#I_ը}z§n}Gd\كR؄fKs{ lJp"*0csP}E 4ij ?vo%svP"ZEn I}B6G/Z6 947U'ݫ-:eIC{{e$}{ c-][ bxe|L;05(s֍cUJ2YJBwQ$npcì?+ US'<\.8  W% f[mC&N]DtU5_ჍV>ZJ Wj8cnёS ۷յj];?!aFHkԄQ۔أR ='=鹅#;@lfyrsl {@(ytt_J 8sn0L:A8X>Ҷ!^Xգҽڵ#WM}No zW_p0g4ѭ;B`YmsUrSLc[ >FB<XcR̈,UI>;x5fى;VL(bM{I8-Ϫ;h;u]zˡ~ЦA<4.e("xh`ꄹJ_M`Tޗ8|M@G/P7ˡƦwL[Zg۠B RY҈f7؃24Q́V cj{_ loNb (N)&خ4]CIˑHI ty𜇇 ` fJVzJJ^PIs]-8DhLr =&N|A2c4?0@׆?dTa&.֊+`.QM䯕D>i gqj۹\"*zpsS:Q7KS<8U~쯺ϫP rEC&CTnIZٔUҪN[hID^9eJ7Ձ1p;,&C/uZn|@Bp#iG6W:Z smC_u~˫":?DHA@0CCm882%j%_Kke#ӗuS'OF '?k5PC߶Ȅ6G&ܥ1?Z|-0Vf4Y)6yۏFS:*ز\5\W)fb>R~B0r6s(,zZvv\F\eV04_,DKPPCc_*WT^ADf0 DV)8yȑ'<}{2#'w?L*IBt{ί n]wito\}3q8 o_7hnX);: BxP0N@XS{}2L'YKa~Q֨zM 152tq'8 dxs,;m,LQ4VhQ~1hPգTV⁵.׬AW1-d st0Qt9W2.v:>=ĚH8{H4rdaV&UM@ʇ"*S@{"Кet>_җ3Ng;ff7|1By‹tr0Yx$CW9oLƿi@l%u?)_NiA881܂4i T4d(d\,lt|/#tʑȎ<*Km(gߍtq jn 8vQu9R `^(P W.:+w1W2L.LX8tћ|:/dlfa/6[ ?k04'K tPlceQ袼CIl^灟t=O/sdMv6kZxx,U &B[j֤^QY0qfzso\hl6MVŬ8 ،ϫ-?1tͱr3n]EC:*~,`M U(m(BTM 7WNv_p\D0Y]KFiλ=G<테@\ynn,_%#f_jpFU"K 0_Y:`;qj}]]ٴo~oL7 %Q*ΆuTie'oT LҔEtp7OBJBNa| Wr4&{fs:GsRY]r4on +9}ȞYaKBjŹ.M'eRdOf|2`]6ziWѹ^^B̍j )Um|PnQzAЧv׿L߯`#~u4[ CZ1T;<`dj- nj^סmeb7{.Wx{{FzYd1{tW=c,se=NXLp<94wH)ʭd3 ۈvT>􀢇끛{A`k\e".' ꇫŮ椗jl).9T tCÄ:ZrK *dXbɗ+{+䀽6~8cS^hoSX&_e+CfúMGGNeEmv'3$)wJ~Z$DVtd1y/-D6u}$sahm&Ol7;Ź`2{kUgjx=z;vm |@-eG%va~*r~ү,U? ZAa4p!Ay(Lگ5p64ݚ+u~F>FQLۛJ)ؘKip%!h"d:?X%8}l^=Z j)`CG!h[ٱ'x PK] 8o =Kq?)XRWpfʣ~ˬ8ѻ(abπpDdTg)`xemCZY_(FY4d=Afo_}ymȳ2r)㋍ ιP%?p;XBV7uGCι73&zDtgm9 ]M1?~qT7jxk)]៣D2c 5bzޛ< ny '%DC {p͍"ys ;0HǏ5/ʅP =L nPcYy~|=# 1\wSoW,/B<}*_ͭAk\6v@cl$%@ ^*xQ)4Gua>सu}NGQqmzidӊ%_\krVAaYB\Y ;lmia2`^-_?K(@." q:B4oEh~_7n t|1ba="lDEn~pJW@'t}}No 1?𮣭+oÓn" ymi:*/<8T;/H-UͦFhr^Y$i. U@niܸ!g w\\tۂS@.4T" w_. Bp{ 2M ԰9sy9wD~Rv;g-ȓ*M9t8!ϧ"w&~aB6Q00w6 1&W/"vH;ZqL[&G,f+c^cC9{F٫Bmݪ&5)#Z>.bJ裯$}nGSwi´UӰ)La 7a] r3i2iBQ So7'Vn٭hKA J4FOycBŲÛJE. f~f~Utd,=*D|zy6]34чoBBqeɣh#ChuHU>./Ϻt"TUykk MLO  1H*G|⌕HFXB5=:Bc&JY*7Rc"G'8\̠%>[5t=ӣRIdU!9&9") Y%܉ǖ'\Z#ǰ bC@%zj"5"$*{ ~I XrǑ"叶w1)'G p%W }-r%P|jsvI?ڀelN` Cq$EajM84Tf;mp£L ,-^|Pcz:]wMQlbNbze$GNJXNq'ُYޚjc} )c-|+Đ7`5cRxC7eT$,AߓYΦ魚ƈ.hƼW&Xi"+qoLFR!khKEPk鶦z:nLD" #9/=|b4s騒$Ix.iH)hW}vou[+'E=2q'S5)VN DgU ,'Q5%6&Kx^41m&15A&w"1Jt,_XgDb$msFV>p3'7 9`QM2%Ef+Q,h)T2!_"$u-=ο;]zphŪ Jf.dTep-Uͬ6!NTJ-S =3^ei[UHiyLֲ gӡ761bpb~bh<=hF!llYGl] NYX%B.[)`ꂨ%n v&: L< uнwFcvNVe\ +q>< zQN0!\mU`2ܝM1B+R >-%4ɬPBf\4Car2#qꑊaRe OdrXarє+ 8Rk IhwH[%R+;a J |0R+I,36#J8ޔT-ׂ)$mPMk*no-Gl-E2`aQRZV7Dʹ*5Pԫ#>4 tfdLp¤s$>3ݒ'oInq3wk_=Rm|yr܄8ҩz9!`,L:jpAzފ"d,i#R1ڊׄ;sԾ=V]VP-nJu$&]ͨUfZ! zvipVB{TRѡ. So "$Wa- ?&`򎰚pUEkUi=&3ŵ>ϜήPrV4BP !1ΈYgP##"r!dzL",x{4%x䀨0 Zoo`ԁෟHT 0EmŪ]Y -'[Pa?1qפ+ܠ8p;guC8uRj8zs}iHtFAx{|(zi&)oAfxO빵w!LF/ V0m':ՃKUW#eZxt?h *`,1)M\n$qXTPkD$fw9 u7ʝ&ʟzq'Z)WfQ>??7ӫ`8EA&⟧~8O>z3r;ֹW_f.ֶց)n?Q 0mԂKZ.dE&HoPDG Mޮ5[(E=EZ=b-SF~jp\N~&\KŠ2թ*?vC !I#bSBhZ[Zd?ж7r׃+[V>2n%Cox  oV^q=*BvDc ȗW-Nܝ2qz1a6t@XH"KJ7t+^TY+M.y]@86U}:D K{c(Y晱Xa'o}Tׇ-Q tx酦 =]p(U t*ͦ*.|?_\a&t~R8:$AeF?+4Jg3̣Я8٠znau˭%G1r>F?Y¨T0U"|뷯-v)o x %q_O<j*ܼbm{bRUvzJnªIE?[XT(_NC=uMk؞ v'O K WEOx!ia|0U_&dh=pD~>0|2OO *~y#{o3yOWE?- h-w1dϮW}o:/ϛS|]>IãP̥M2*):h(XH=Rq-'m1E!}i#A`^q\#QOCb̦ ZxWL*~G\V6CN{Z19&zV&FFx= ^ &}Ri,YX5ᶼdQ Pp9MUwmw(͉ ]RS».]QypxU,~c¥jǷEDܶ=)¾8Vzٗ>Ǩ "[:Q&ݺS#؞Rpەݺ[wyc v7+4QRZ D>GkGެ Y<6w3bΚR:  (煣"gxpϮ-rn_ыcy jA-0%VܴК#_Cٰ+[]C}?_Y cOח#Q6Wς] )e𐺶^eig<5 5_z+wڷVS,~RO*l:x:(n!cIGd!NBVF}_) }2$6Q/J ;w[ l:O(ߪޙYMr]y4MqzwK mIHORqɩ@r'&/ɋ4e'?aڃr.ҁ7cr5ޅ?@p|Ή*Yh)pae "cq3d2!hՍkA;!9PI 1@Lnl/&R*$W CZ,n4"ǦNjTQ |5.7wfdru 4#,!F0*^TD_wAI,'< :Ou^?&(-԰&_, LY&4r,e_qI/|\ŧOdo'#_B./ެX,(b/'7Xx'W^,`Ĝ &c)gӋQV- _,n5I7>HCΆƫ?OÕ/f~ʝf~ܼsTfR/s1u,w>3"(7cL/OIN&J/z늨^P[4[iA9_JxAҰ2CaaӞGZ< >#app"DvFoK^k_.*:;Bevy7*Eqo,b(Ţ`@[ٿ^W㡧s{?M]WPAq_NEqVF[q$qFE6֒8U ҄D2R}X! :$Hے5e@/ϾD"6V+ :C14"f1AZ̦1وM{(Q6WA/)U]>l/vp(%JiO"aVJg)np0b8IN5`_I)2sQ̐&!\ߧJN'qޮHzRy:k+P٧,NM/6P鸌Zs7g^\J{6_ȋE}c৔< U,1&\~iXmߍT7'`Z#X騛7;ɞQvϥ)+-ea^K* ] poXX~՞]A2~o~nz΁N/ 4ۡѧjvuV|7?8 [hWWn?f /4#Ixᶛ_(\|sL>85y}16's eq*aky!0hC74_3/ƲpKl~zIL0uhV4]6=1Qsb֨xTf6/Χҽ/^ s?}[hۅLyPR^ӓUn3 >=dJ?6"6MR@>uvU'5#}sd5 zOk߲| z@˫D>f~Z sicHxfD'=TΈX)!blO,?طK_%r?O+cuk'@hRPp)B<}ƌƎ&2@ sQFXp4OC+=ymۤg9ϷA}co/ nke [yQ?M@6{bTQBHhjQ 6 zR1cljS&N9I1y;^zm}[+Ffcss6{1g`|}٥9ou3X\znYouV,Wou\ͮcRm)d{:9?M4[+( }'MUݔC&gUTsZuET`tm}778z[4n`QVaWp?r :c]ax_$)&-2NϕjZK?%%DX|?KEƛ8Or0ZFG0Lrt]l%L5Q#(Lȹy 8WIo8s6`Nu[ ZdhMخ=a;"h-b& UwҿOU܍ԓؾ]ŎQ>~U] xA;C%ɷ[܄'ݿ]ؚ=cRVp7|b8,` ֐LQn9WRc8>sџʲ0%qwԩeEWB 3;qǸT0#Q`A d K!{@3#EҤL#eFZTk40)g//?,n~=;Dݧf1P?V[3&mb@.p2 ~+[{|%eN~S8B].d^jw,Fu)itL0񂓽2 n1cNPǮPiP޽Ѭk;X>Tѡ?S2[{-Z78s]>=-fo<Ycڸ:/H낟8B3s)K RF@\:SF 88,eɗ3EC($GZc}zvSz,V,>~ $˾cFC7 [".0T*߆rs]ZץF.K܂.87{GZcwİ掴p3؀gw1dyR#q{1)Х ntN7D?¹,)7R>]DoJy/RBo |4+JP˒,s*8-X`Z[dq y;;.<9?)k.tߦ?\Du,5aZLc)/t1-=QK%x)q\8<7x5ssK-hjv9qi1sKG 1w-Fb_i^vM؝&.P!ĜTj ,6YB?ճ~y`̘%9Kg ,FYB?KhIks)sowu]>zוlRLY/,"hKL9{@_aְX8ozi ;v?V塅c5䵱>!%DBNpG}2(MI1eK5lf<`qc.a mpP6AM8ʶn4eMI^݀=t\CQ!,aӧeH b#!D.KJeTQS9 7ʡ\(#@sV> mܶ^yCÐխ/?yh)fAɎNM&i\ k e:86Y1mLAƺ(J\ \F"FE,stG!@<+Snay_,rWjnɡL!NIhtKbP@U$}`CxL\T(iZ:plx"-2*|G_"r'N (fp=% L+}i65jW4?w`<`L=f84,'-2*GWN/Ç5^"cƫ1w4W" t{jPuHURH!Ԛ9IޒS\8R{P-f\M8ުAZ;}# 𡈱@@-QG,"yds1un3z=IJy$d.tVS=\gQ= S)ir7a WgůX夌˖$KZ  ^k&(d~^Fu xoyOs>俛*\E`Q ]Fd od)D`p=<*GU!_e|U^#ԦE|m@&gL!?ux¢D,`[KZNƞ,G(*|VCpx!IBiJº 9P Q?XP}U>ta gg`QK=1o4EMXU1~oYȘqt5}z3ReT+ :˽"iZ Rg|P.f`M8ʝT}vpHuǍ^Y$TWIAo@!dT:GTGUY.iRsEXpzr\ E\{/?vI>>gx 9Wp^.Q)| ]ڐh2#>PQ0O3p̈u3vG<)16PF]1GKY!h0L2Ƒ]] '\]P% 1^"]9;M3|ww#?=ch#&l~'~v3to}vεa5Inu3ZZUVQ>j mIEկؙ 'UwS|KilJPN6 V|绸NC,Ոإk\cT&x]Io 0K@x~FL?7Vz41Cj'Xi: I*UYAgDžKG5+Te;t,`<|'^CR 2Ƹ;eTczEWbԦ[g>ݜ ޵F!g 𱻼t;faӮ3 2[&Png{uk6i5IA -א pZ>.x9Kƌ(򾪖b"4]UZx7Ԟ@aV_AXa{ٔL9AnZlpo8Iwo;0vvJDd8d``xe@*zn)#'ܫvv(j Z5jc֍ڶ5-nZ (:bk~f@y_O yi1bjH_;Q3α+d*$MhS"cƎ#vzivج_?WOq%( kBԧĈz6Q8 yB"I'yxW0ȘJ@޳8n#W}.{H⣁n/a7Afrn]P$HY)ٲڶKvVMLwEUQb=V|8OJ5&),Kx謻ksOA%-1$F,L1Θ+3 ΜU[;xr?20f~%&&DŽB'oܮOģ:5:x-My] AGpq [( S!N'Dԃ))f3S \x}〾q$NME¦ Sx}S<٣o -'FIP', )Bmx7?2bX. zp"`?#]noPMm) =&xpI .T\n-fJ"bCӻSix* *6m ?# SpXH>ɣ=n 1e^NܔޘfS3!spnQ!u@O̼3,؛$2֩gnS{$?Z) y ?cUL(ͭׄY\b$Fd=yk/ϥ愾PeAhzxw_4ՄиǤIդ@NJ&Nuu4ܬ9>X|>80fTkfPl2d Ek畹(SFfJ2jxHўcs ? /w+OИP'޸):n yHs" IdgdgfS> I2:>< #HsYR8hjKΈR4xr9Krp)EV2 Pփc%n<y1>Y6`qZҌ!.U\dYIlp5 sI>!O:MЬ OU1DlgY?j\RᰩNӨC, xϤH@EI@bzV(vL=R trA$- v3=#\zpLJ;֟~'<0KL:<1ϔJUB3$x ?IAdB(̸^B:DKyXg;&\̫=8vk< \ +)5&w hIT[udW Ϥ8`,(uAY{ O8%4ZjՈȶVuEA?m)Pme^2x6ŎϕnR\xm_ zpLx. 8b2Lh)p6==rMzAG,8sK94~1"INj4 A1$'eZz3~ ?yֺ0/\w8 G?DԔIq FN' m5̈́"3=ȣ]W~&uՇQt_x& O3z&z[#6k>4 `96^]LӫH%e *ͼϧdE3.Gܾo(>*4:5ȥ E,dye s68Z~.K+]{ͩtU'ܿXp51'&NP} :"DoN#:'7\3nP tsdNgīľ "sI\9@*zpLˏx,ـj( jfRBaDiCxߡ, JxJ48ɡ`h;vBGrGM⣫ m:+ $ǬɄ$NU2]1IBҩܠA6yxӳ1vHSoރc:'LD%_^"x(kAB/MkP֊T38tF78c uޒvt:j|lLo;Ą\3,Uxey;+w*>o: ޾c2hф!^Z2# 1p-58 ue>P*3fD.MյA x`\yY zQSxm8a]-ah3:Fj9[lo?q!goΣVϦ Q__כK: đl cascRpJ_`;SR0z>VK6ضBz) 2 CrR py@>` RVQ.LZ5dlV> ( 1.qR.d-#*/\k)s Qh:b@o8ǀSp(ys$W<% 8l.*ûKw$j5_[] `M3>sE՛E֕?=z~]~n?@ßʻtշ wuE`Jv-VGw!jYe~Aq^; ς} G~KG_h}e2|SDy}?k9Jw'}ƴ@ǹ?",9׳ퟧ_Z\-쯫_tțqͻz1=$u]{9H^߫C=]5z`W^U׎x!YfƑׄ$O2'3Rcik%*͈Lfԛv;/\qOsmEQ?ϗ2/|f-Kjbe[*c͞>\K:˥/OˏF-u.$\l~V6Y5b.Qoy}orFArky;ھ̓W~EUq}cs_coewE3L %wj{l(5oun޹w_Wr?eH]ŝwX#\wk}[fOQl[;0qc4-`qr}۸/w0G揳fՋ{yGM,4{h<-@}{95 㝜?APzAS 6.B\C?fQWxЂĬo`6ѿ۸Gy9F}2᭗O6r\9xw6U^ 6r DZKpK$yBHΧ Q{Y~ ;K'=Aǵlq G!Z8z|HU1=pfyQ?bޑaDMfKzVz^ϡ{&y4 :(YW ` mq Ahg层zQ `(Gm\Jo2{:[v*{#J%IDKb1~ImIY̝0! VD.6۸1(I$׌9pM R!'49>,rE ` /ÿx*:c7bn? { \5Z$0$EL},ϰ"jT`@o|h*Qm\̮D8sjչ%i$uhfkYAX# 5V{=+#qKGƼ;s,j !Z&96.!-II\ZBB)- a? B{)x DLzEk)sq l`lqۮICW/-7̄`k ~ pMJx,CvrfFffpJ/~tGrw13lQ[/bN#16"0)Q- v,C^K(<*YJ3{ ־>>'RN۩4|]<{#{)U ҈'rӎSJg z~9>Uɿv A3nQGwFtp &+&.t&B?3'y/#; Vfd&7۸ʥbg4f=`LZ[jW%L2~;o=U"T &Cg-tT`żI57Ҕ!gNLhRH.5oo6}\f]͊`лrЅٟbkս"*^2~z?Prp"$adnRp??RJoCK3.=PRu}3"G aͮgo劖R:xo\5w&zf[xK Rl' IksSk6]J#7 igcVIBzO;J&Ⱥ 4I4@%M/VGw) \íL$SLJUoi|407bKJ2Ϯ9k -ꢚp;0<ֽ(i,+XԻ@mFGODuJX$}Vf­Lz|ɂ`"IA&T# jVKm\N*+!JsSPO6]83)N!F=_5ʳH1=ݥ B'p S&O 5M(6lYj8(gzuj71M:iwp  -eQX@G׋Ƿ]Yo#G+@cw0=300cޗ #jqD4Y{U$GJ6Ä̌#"2AW%Jb:, |˖hK]|y1Du0 2O!E橢`m]% V4!=9 P33ֻ46( EHױ9SO^U^U\x Kg`T\H\Ivcױ[.lw")ď6E}s0t \.On奼h(x"fEv]Z>vu!NG >3ˡ:&{H=< "HtS@IL$h<:4խ0LӸuRIŀ2 l 1]"S {jN#;܁KLzF[rH/8%ֲ1KkU%Oy\b>msKb f `ˮ..~6Xv”ya !fHDB/\(88㒨l_i&gcgKȲ aP6)}-@h?$bn֕Qڷ+Zci,t/`U˃VviYh}3?Q%v;@歟MR]b;B^f4?ᴮ&]v(b`H ;e~^l-^kuZhqڲ1YYP|*h~ZhC]A$e?΁4̫1 I-QKQ˷5eFeqذ1DgvQn1# 0Y*Y0ψR5R+PKn4 .qn3Br~r3xJRS5&SFc/><); X(-?q1,F)),OV@GKBÎOu$+崼XD1Si19-wJkyƈjXZND#="F6n"63X8j֥tNEآY+N^9G:/gviä0ȚSԭS'qK\𻗋Q Yn92S:>`Ԭp 袶jhs.-)BF?C|oxA9iU̷~|LgiMU`Qe4e&2WrW)n|xB^W诫C|T1ecaJY]fRǹQJ --i ^2UKwsCXxQ}~Dezfi9ٕ6D)YMi)XrTu gD9}l9fq`js'ˏvl1D)!VvR7Gθ{[Wjh{V(~RR品0rpJG spd]-31"1%M(ר hNLŗ?^OPm:wBq6t T;'XSc[=*Yn̪w u|l O2j9i˽Ӟ]ZnŷJJj+ PpN^ rIA1N+UPۅ-&k2sWa P(9^ί%RgcJ]wUCʥhiQ=ǓȊW|uuHM_Fݔ:[a#/D?]^5J8i=?]>EwϿpEWR}nk(лMm܎C+Y$i,uJaL; rT_шogrҿ-cIq0N%?fhçg?vQy6l{9ShNƟc8F4xˮd*ɪ_7}]?.t|kb*Ke|Q &7~C4W"5oG?F0e7.gӞ!Yo o/7?4IEߴۭ6n/x(Ļ˙;#uz_+K|K!p+H0J1*`q Kmդ6cV}f}agEYb>vY8IZ*Tj</anOo'^qxGn'a4#GYrІUjKGɽ]|f45 roNplnQ{Mִ]ONJfl.R8v{Vڗh{?kߩh jem` orƝ]\.niDm֕!M^/Rz<0ןŎ1"k r|:}/ +_`iΐǭlvv+a!b|*^hX? wu}!XS;5%F3,bx;ZYN&0Fk Moo6`;#"nٲz|{yIz?(.Aϛ^(GoJ-T"Vi?XvV+W 28 +* q[EI[Þ !>c? ,wQp)xri*$bJ"rHzy_yCD߰1D Do:KZdK.7XC8]YNlAWjIsYtR)`qf3lQZzL?ClRDVzCr,f4iD7_ yL/hE,GE0LqRPA I:)W u~6u7峒{gn-іxn?HTr&eTX) G,h"UOWesZ^yJSV;#`A~{TPA?-ؙAa93Иty@£(XTa/l< 畻Oku9 Cx}ʗ*XBE0X^#pbacS2`mf9n)pUG#1[m+ *,̬1@B FoK诋!x[6x  'T YWn>`-uõ!`D"(:`;RP`gc%hadBO. +l$-x9UOȍvT{tk30 SW0ra8F )~^ʂWFRP(j\UJI(p*v)rsnR- -%4ɣqQD"yX lM`i}jUus+!>*M$bt.aO)=˹A,^3.U+TJt ; t)4YI :3o gJ ;de ۵h dv?/V S rfn;EX\s:` SdqLݦ*^$x!rqP|;Q $՟P/.Q NaSrk帱2 {gǓhE: |TcTC|TD7B儱 Dpޘ 9a3eT{z(fè'6r2'"^S*$ {+!Qi 72"> >ޘW!$$N;bs+&W)IL+ ,* Ax2Oݝn!>&!1!Y}TڦmL`8B#QLuqJQ!%Fk6=uRF]R1?R(vF<_ig^1"5úu,41.`dcPX=NiY Rm iy,%ZT\rPr#NuS;䬣W`*1`' o5*3!eLX'XbNUʘYJr.<Ş8b־)*0^W?Ѿr֣s/iӫxi-a] biAՅ\r8㾁X?_Ua Rqi 0FV6(@{bRAe,4d7M芗ng`Ң`T$X@8,Eu.e8 %Dteanv275D` iMq@AY)1ky=1+2+]@AI:񍯼H3Vr .;]kde.hkc=Møu8Zi:rp؈f+C/s50nuf3]}'R/wrk~f!'o>}a( qh YhT5UqDKNaJfd!4zyz Lw!ɂaYޓrG۠VO#0}fqG!E\ HQaEäL)ר |eB(^ݢi?`A78\/qqX^. ۴ H f!ԷKlyc/qr\ .5e#ıO"d2[`/[1)8&.;\t?NB1Jt0Z5RbVDx%т$%=NIh ؂:L?ĺpOSlx ]mݿ.}QO`d_O@o??}XY3Fe= 50)KL#Dvu/|5cxl/WnrfGhixB}C^ %(^6 rTQAԩ_G'5((a1:z8J' [C"s&aMWbAL`hŪ3+UN~)ʭ?R/hk9>oAr=Q y24۝ͣTy x5lMq |Dz lvdvv 0kGNamqrW<C]@0;Z-r !#$UUCkzA1Y˴&c6DL, 6EK3&HAYWZ#Q!R{U _F1첊sչNas6ۆn+x"[Jd{Px)U58LjX-8d|jSUxO9.}3/{ hڱ4Ne1j$*+^5?aaj/׫iě| 6Qb҇Up L 1A)FQ),U]JR`gX\M{=a˳ӰwqDEmF7%,^JAuiQ Rpٿ(Iy*J;Aݣ뿚L1vtVbشM?mrՎ[zx¦f`5T9J8kI'ʊ3]o6Wr"R|PGöE;r;łEHs}eqbkIn7-0qsğ9VGx 'dJkHk~2_e|/s*@|dQorQk_Y~ _:ռ?xc+`Ò\Xc!16=ߕ&oSUX4O3Z_>eR<)&A֞MV3&!TM&)xj #h}9x LG5lm< 0<#< &~eN F/k7Ѕ &Y,8 (T`jB:=!_2݇AYQNHkGbpl8TpA(&SeR`x( oiUJ&AA(7ի #Ä@Fgpn0͐勌3[six8{6o殷n&(?#18[CfcIQkez H N銲!jCH N8,VX*Hb$X pN$ڡut͢BJx@s.hl8Ŋans+=D(pP{P)8S"'uVgUD2¹Rr YLRp9PZzo&~ "&AZfY޺Tx4[(1I)3gt H :QɊPshce̲RN<҂}畏#guwFhiudY Y&iR䊢 b<~-O9/%4p<2~8Fl FNIiIIO_gU~`-$~b\Eêd U"ruİD(lͅ༃-XGMKED.|s(s }Obd|?n3"eH "$̇q^I)E9 ؖS*4!16 ??4]hGbp$$RP |$2񠊎",#fls42J,#>% ~|Ƽn;]U?*c-A|Ib$~2ax1hl ]Y(HI)X|sxiP&$FH0sTkK˻2(2)+*H)AP9)58# PO QCM]K,MK{Y%F+zFqFs 9-8*`aB 'JE4-$FExT4 !y>*FįC$ɮϮGɂ?-$/"}H1E9Qr4#18j({ W]KC(Z rwu]I&%FQ)J ?HDNwߋFƫTbdgb"H}GN\LM{nDDSwα A%@TJD3c`RmN C1;3v?² >#18y1uAh S IN<loE3]PiLD9>9 E"d սQ%-ʬQ̌hucMQ5#2RZ= d@N F̧2-ռd?'Ǔa"#da3r>P-]S2zJ71xo}~qPa=V' K (Eb& 0(qЃdrOeVL3qmp 5ċWEDBOV$"H$wcJИ T\զ{'%BN_UvD*F]T6/‚ exV)+vDf:gbEY%L)aB kdZ1Q5D(h~Xb}塵cz ;|! |Lq\["a6Ӆ<K}تeTK3#-8*yTF/AzĠD(|>6U :JjF)w0/J:-H=OV"FzhqD$ %hpq 2x$W7#bx$'2HH /'gM-gʢ2AQxK0}#~7Aڻ*jqS.e] O]eK]q,hiȈ>[!}Qxd%ږ m9'ZQ_1ᒬ` ߝ.^fC_76C<]cAqؐ"a0E䈍|8VknIH~xEhj벹[6=#I 93#\="'N Y݌dȤvN#hGJ=VH>1$21O2úȍa/s5`ρXJCA񎉖Ϡ^S슳HU͛YtO xeg1E%U<#b4?jBsL 0?bT9)keF,H3SDႰRThvb0􃱂nMtEw: `i$E`+S"C(\H NCjD` .͘@3MiLYVwB 5#18̵EAa+8nJ8%NIp᪩#KH:aGbp"f)`%Vn9pğC>£ i 0+Avh5ReTRv-"-BkO|?= 9#-8|T0J2/s?dͩLpI9FܡA^ ];!lxĝ#S4BpOzSyIVޡNPKi-˪cf5A~g$F)Bֶ] )ŀ$SihzeNbF?1x^\RkJdYNefIz<+$Gbp^fpPkJ.}dJA$I))FS0NjnTGbp"_vJ!~eT$SX'y΄BB/֊@ NӜ>z#-8"ޟ$x$'A'31zGbph9Fݏ9As<x$M?=(;U\!D{KЛL҅gD"O! {cYFȊee{AIG #mll 'ʖHf(|%*bzxoN3ȷATr 3!v#tmY jk8.XP6wi8U`%FZl,i:Y%F+BU ifzd W"CvD{VP<#Mi"~ݛ o"Yק\ca H&D6x /K37A7#18*FdoFxAEb%W|nHh̺ji8Xf0'OKw>w, X>m7/Ӈ}8^kʝm-uW?p2 r57?ˢ@Xh|UjWGJ^1>xa V_Vw?.OV;t߀um_xWA]JOp臏>5x|~WwAn'痏dHF8"m?n~8K6Sn4^Ob~]%[b^wo.b/ h_}Cay}g7V?'fH_Jyxd\Ta=ׄZRb\f0R[ Joe4)i$y TDجwng#r^nu V9lAEn^[80WUm? RʀxfU ?nm={ g7uRF0of r~7Z__zqSo{쿯$\v}k:o֟zUی}}Zꃡ|?xWx!rz^bs)]QpIm],,ԍ|4ne@>|4='[3>~;5F*Wj/WpoDbvО}~/9z!$ۿ=B*R̽/1 1| & V'0W+nw zڬog[j϶enH8z$a5PÞ]GVC3/8꽤жC=Ypv{BH=at?ػ6dU'1֤>^ٻ1d~ DrIɉw3/U?;I x͹r+ח'+cxܥ(0ʢ5 Szy1e8_sPRT_5iRDȋj/X}Tu(n0S9h1l\<-*$w1*tPf~޽X(t$ M/)7|ރ>h]o0tjF%|$0NlP/Il oAFyu48 ۦR)YJ[L ƁOCrIn%sFfTXT`#{wQC81{1uFz~E(W?T+v:]u'>VX~ʵR!P†OS;/^S"\<2N(0;;ή϶y%u%Ej*lt,p2Ge&zqf]!DQJqTQ&#"(ylVJH,# *U$GYvx,\;Mᳺt2:57fh9{~{rS{Em [le[bnw{ `x݈( lĵ!Hf]<(SA$qp!Z ҽ8DlN9۞Bi0g)1=N9mHʈ$pJ`k`]D1k$0BN#)q\ ?&#)4Y 8D\Z'uɕiPF*JA0HÛݓ?BR Z"B "r5נE)ϕQ84b!)V7#.&Ob1傹X> L7Szt( .EEg$V{a7V1b"iH}>w*=/euquK\aO醼tҭ~ \ehoIgd'iMH isA= zL΂ i_^ 6‘a.&1U;݁]=v5}Ň/Mx)clm8lg5d8e;b;N\X fS-y*84%Za:t"a*¥& S&pZ]5Rʃp-1Kȳ8s =_yjK]w/a W덑%4 Pfhm8l؎HsC@eŪLܢF9=W]44$9VD&GAgNk(-4 <-2::|sY,#DJNtJEd403,*͸S#$"z' Lx/wSy, SC/jPIX] k+V1|Sc(Rie*r} +i01dOokkj#H\RDSmA(IW10NP6iA#&~ Yn][lIk`ydDzl+q٠k8/U,RR|\~1z ?]y}_,:x͠xy_B17zFk&)gX!]/.#P\b"E%ScǑ3F"ҟ')(l8Nuol(85iatH<\+or+m>C> ߖqWX|X=z7/k11*+(zlr\&Ḫ-rթ>H4s~\*\~(4^/S4\=|:WzEۚG?\`%qOChSOՅcO7t-7vSg >ƥ~WɤN.=[9<_39^ $7ij,]9_2bRkRǸqu߯65 UcQpg`o@}|}ۗw_~'~oRQ*!^x\'?ܠkUߢkXEM)o0ɲH=t- J??\r_ gn =MU/컳"JvZ9Ʌ_|2/öݼrT<*IY-OeRlM2Ryr>4 a;Q*X:ZcFQXB!d,0y>)g.zh2ߚ?šP}LD師֜v&|g6zOvA:$/UGLprCR+瞅h"HRJFoe:ɞ3X;'t*Y66}v;]{_[\j>5BIVW.M: 5@@nvaoֹে_s䊜uEeD.Z}s^?h_ /#vѝďn;?z,QV(Z3TөMEjmB3R{Hm!d~H<%T8 ySMKI SٙYYh6.g-fM;6ҌFA/o KK32*1r+r6X3+|s,WܚfUq[g5Dv k(xSXkn^e2&j*0Q}@qBη k;ۥI/f>d띆}쎇=; \;K˦N說}gDx7:?87)?*̇\2-sq\2-oIjEr r\,W8r\,W8=$)Q:W8r\,W8r\,W89s\,pr*A†OS;/S/Wװ1gAVrm\[,˵rm\[,˵rmGT[`k r\ -H˗:@Z. =m"`ҲwNqK ?q‘,zAS Jw5Vf:BT0`Yx8aLra1y@.K=b9O}09O}SS<9O}SD<<9O}A9O}S?I D _ o@yBQZ =&R8"1wl|@ήj,O#@\i1RH FY""bx<ÂdF3N/ #$ SL-yA)uab /:Kd|:t_iz⾦HaҤ) *Xc<8acf&0_ͽ0%)Kp'Iy^6ҙW9*X0R-8 $3*%!GcWvB@`]@KeVѸlUD *Gd4&jI!F;pϾQ,W5(t%LŶI*\XJ\K(:ȥC0S^=r_atGEm2.3:/ٕ8gZxȭ_h Ad7v۠EmXXYR-)3z3* h!?<3lǤٲy4D&i\F 1#F т EZQgpS"+}Avvi>XflfV=]h%0lIZKp~o^ j7sFi^܄gU_9|[![w-Up5j4Zr\롿6#;CUs@sQ7Gݪ a"_]NMNbX'Yo&8 se]UF~ڳsɪK@ ~߽&Y$InfW(ٖ8k\fv%=D;ͼ]뿳rZ?X˘XBH-#[ !f{x]w6ƪyyP uK762G,:*x@NpHN&C{P(^cB1y Aie# .JgXP aLb.kNb]좖1؀ؾe)-KdiFZP |7%mN,0--b҂#uf 2| %+di.2Ľ)5kJtAx#xiG$$QAkۑHaCP(^#߅SOD ȼ+| %3r_ '8hG,6ѦGf*Q<JHHǀeTCPJ7p8x8 ` d$0CP(`K/fȬ($PR: *DȜN  kiGy9ħڂ&=.=(Fe4ALf(./J`PaP(1utu6b2*].AV*m0P[)hTSAAa=hݏEuWe2Cyt/V5\:Pt@y&5q$iXJTXCx1.^3d EELY&wS(r)WQ#cEJ-Q`;>J7F^>̪VmibxXՇ#5w,P<8x# ˕= bxFNs=(O, }h%hg3 : n}(8--F:ݴ,7VYʅ 2߇B \MBPr1SscyJxń氠׊+qMDy}(/5) YރH0-;{)5]{P(^)V:r"0Έ%ow"w烥UG؇B w8d́W)NH P!qޑهB}Ak)5R%*umP<7s^`f'x3@/At=` f*F{Է?n!+cd ̴{ykopwjr.Ї}4k?][B5nM窕w j.WXc~y]䳕exXc;3&w})o4 }f׹BXHȍe%2),ŏeoh. >\0P\2em\緒=2Vq<H$W~6O,ȩ??H < v8EtbZޒ"gӾi6W|ӳ4{e.ҺI\ 8D'Sq~Ґyɋ'`^= >rցeIǎG'>0̫fSp%IEG"1WT? 1$ S.Ha'Mlvxբm0?l]Y-n'W<(mMĽ$FE2. 4udCnx3['q&1 @H,y4Uܫ4'mF+BHD(ѹCf(`wq^%mC{Hjb`7ka3g<(mT!F;ad:@B )he)^V8r2M} ubb y͂* >*]w$5PQ RM{ (!QÌ k5fRbRP~-!PVx x%K0 MW4w#>ˈ_FQc\Vo'|H&3o'yf$9gC{2'|cI>ʦQ^fF^Uv`&U^27kMHVIX`HVyw(<}LA\9S1.M%(A8Z\qzXx{w i G W0nc6:!"nރckV"=|[kˋ} lr>\NuwS1Ee 9:5`F5 mu yzo'^{sy}jypvnbי a꒐pܼnf~krkbmÓix:^\:]'@I, U b~Mmku~+̎Eժۼ. oMm_]&nۃ$E$~ALgFQ!*2QyƤ$(/6™.Igy2FI=I_{aP->y+Mx}: RBnjx~oִ!r[~c4"aXs2jxL '#NjU"M":\ɾ;ñBC1Ѐ{/30=Y4li$sPk|GЭQM⛽,0U'EN؝IU<(dpF]`lkUClRr6CԷ/AX3]L!1^vI("џ= ^dBɢM ?NWQA`3j_:~։pWHzq19`|[寉Q2{$i[/I> HhWuDB"r߷&P~2r ET M$^42N3!gvHtS授:ܫiwwmH_eohGs; ]nO[c=?[owK-C8YdU*fjhp"^:#J%2 Ȝ\2[J$U݂D8ABBTc8A7 ! f% L@)J 0Nj8i[ ߌ\.]31}j)~u??k3bVIdkz;BvPXyoǾ  =[vU=:: : 6rYi8rk"@N4F*TyIj{pԃӂ#C8"p ъ 4\ @ \9kANqRi +hz\=Q glXWͣ O2kE^VW#y,,&sUҰ),\kRكӶ֪Tz'z3,[zU35m޽&x:{f^ɲ{hY7Y^5EqXTGZ'AZL Ɇڑ,[ HIky"r .j(@ϏXV63y*Z~q S @sAHO-P)7NK C X#DhFI a 0͌\j"xX0Y>'m礋w<ЬO1q0N+mN8B*E$Q_% E R Pݾ!K)?z2%Ѩ3@8t@ϑ `PGBҨ};)@ۉODذE Lp8aG@;v^)/*fJ@V.8{4i !,3bYʤ*. LjW~A~z4kƿ][ Q0?~i@i܄8Un<$}G:eY4u(FR5J47֛ces|IJ3:]5;ySc @9- Wa\H{efU~W|~:L3\ʅ\^q;/0;s{&1*V5 1Yxة ̗Z0"ZF#u̦Q6q8Ĵ!::Iր"Qf~](|Kbۥ-1]n\=Ŭ;RNgML1]j*)5QFh֡ anAJ/T:CQR$ڮC=CF:&pj+H6cRG`D6~X~ڦlڏ@Ҏܜf2AV3܎&$sDB]Q?ߔyd4 2 Ikp Zh>eJ=GF!` gM2u8rrư]ε螗nI7m}lTFQ}lTFQ?Bl:06H>WGY}>WGZ- -y uQ )ӄy#O4$> z8Fl '$6):ve>?bVO0s9F=@hsV5Jk qy`boZId\oW0F|/ P~JR3L:Kq^Rve.XYR rycKҙga`{3TdmF 9W{kMOfYzΟU&r'iiڛCviϴgL}>vL}LxAk$֫yUaG@73sMQ#3`c0E$tXKK {{HVK2{?7|Ž稆Q]NN{׷EO6xZW=;c8|2aptoU]Κ;tuFDə˨r5QFhZ{ūd@t-mq&Ow fvvAM%WNIӳE]lY76>cI–ۿOSY󋵇_G;(5s$S`E!#{i"+i\vaTDI\;:Hmo];:G݀VynjWb#_if%4-ޱJ?6Jz1827n|]]WZ%}F!\k:[X qQm߆ja4t%ʼn%a>P:iM_1ZņYjkƙMlV/|5ib=eI Xr%ЂB;hƌ1P@jrH)G--|&d ;jbTCa 5X<6#k $vԫՄVu<yCS qbE)xSt0; `v@;t?eݵa $OY+--1Ąu/BC0\Hnu.C1Ed [cEP*]@% PQd6[cJqŖ;*\3Hof7RYi]<#FKmJ BPzE\@ X<]0@w3!6*MX.Y 7_S$*or@2M\br%bI} 64d|+MfX>-S@z*i\޺ʄHFƶIo/g$b#B.gTz+therhC7A}d~}`cL Ho5sf:0+ FR"f '~4ܮ9䗇YVL*,'@YOE=˺AH^b6ui& YZ|:=-&:]9g;rGyau1Ȳ^J>JkkӰ|27 ~%i^ߓ=o&~N3 OA}pw~WuSY泿(%&ZY؎$Y倮qS]]cAt͚|«+rI϶Gogf,Un04ۻ}?L2W]UgEVZ3X؁]|%8cE&3o"/Tq,Wkόm&Bbҥ)Y(@/ ղJU/vGhoU#H2LK XrDMcƠP#$̈L+sNN퍤 dK8tP:Xn#u0˘r+%Bfk0X畡9BpCFuR7{|KhhlMl8@e7S~Z{v>R9yW^֕U} vRwٶ?HЙ^&zh:B|"H4.2:!)#&N:A*꿗";P,g" Cx +x0 "sr`>lR+̓Tv $RX 1Hk -e@Xl 1^gZx6_,Xjd`0i <eȑ-/߯H[ G"%BVdj980RUeZw};s5(W5;2=G=0e}t'_ݻUQbÓH\0έf[(I{}-v[W=:T*:ڙafw8 3 5QH8 &߂G=8zZpĈ$ץP4 5yI,”8. !0*8’E*L<iX{))qJ7D b΂wDðqepZw#Љ 7(T݋Եr^zQI˓4)iz1 B+N² 6 X`N1sD9 me. qh)aordW/7\zČ/a_x h 5$-!t۫!A6^́f\GlŁx+~#ћ0 ;Gv9&r6ȹCM:2,"1rbrX+/B"IFZ9!{+/+F y@ q-Dyݗh([m{D 0\ >,5i:Rj$/M+˜Cqh¥ ST [)L^!7 nG{zԣ;uʪu}{kNo/t p^^~U2x/7"Pd] &ZV728}֤}@ @؝PO'Q~2l֍5|z^)b'xL~#F07 f~dlj<,V>$s_шlAoIO{DF}tIQ1%هې=[ro]8. v,\+&}?z}ٿ t#I+gs EYdW|y}toMu{kxsE71_o1+Jx_0U8Lt4@mw:[l=@:vpzF'O &F)W7qw*+ûsL9i]g-+ٯ5a7*('("2c>jy #>``6Qԕ6K-n.diQzwܡBn} fźSA)S_ϗݝyه&|Ά$=wUpY(MAMg=Xv7EG˚Ly^Ug=wqb٧(͚OF#T˄K.@jd״q.)H|v(*R 0;}I ]٣}&\@g[fY3.sYkgά~xE\/@ɡ"up 'ś44a{M|XP,с6"~\tVzR~&FitȾx,3[LÑk$P!-6guԻmgo' x=5R>5mL]$Z#(k>~d҉8mBhx U[_`1sXH &΅. XOcIZ;=ic<u{>%wPC^R?S18rZA|$heƹ YWJxn\'mG9lϞ݌<EYt7JٰX.PtV3v4&O&"^$|֬h>(yی/6}D*`ҵۛ.$(6ݧ9W, :t⋽>T 2{Uumk*ӭ-2wT.M_䢒w^buMuƲ/v0we[ᾞ>3o do&SPgl;[;Y>r0YtI3k0a"#HߧVc=`>רc&$7Ӣ ,?tֶ8rP;5l&8YփhxGof_Z}9_%U/zC姙9v&?z6 yZ(KHlz4uҷ~٭914ʮ&agi[e5\YG)k /K}N)mUo6p;b;f m0@9%Y%'NvR!n hd{b_$m]p殓vPMD:Rb)8F2Rµe8zJK H'?'nibǵ6_zկ}.Ibl?dWk:Ðᖝ{2r .殔лX}\6K&4JGIߘp<0fwMݒ5"C@X9ǝcފ$(C])xMGsd$Ʒd$nH3Р{ uD^G*@HFǤ"9̰4N+ t W< |;u=Dj6k^D3Ƚ;qe^*-=!݁F"#X+5'vzi,d!*A†OS;/^S"\<2pp (|]HVhjV~tcaz1 QB10(?XqpFL`]D} G|y023XkbH+UDb`gL˷3O/>xȮ^oh7ϣi>\zۭ `:ȕk5sa5C%k[bI]3-k7h} ̹[t|btis4e'míYjX9h34ʂeQ2KIl4-?_@7|~xW߾yz^+8dEd ]A^xBӴUT&M6GdxvUCia,cT[8#6WS(uy׳Mf<l&xģpz~6,I)F-InW칉/T"tGh b u1o|;zq?E>RNIO$9a.T0%"Tlte5fT05,AJ흤ǎ6t^R,/1tɽ~Z5B]`".0^|k{|gI<鉛:HPBך"q G-7(2zYq+(dN+:iߍHhhM&v q_wkKxO4l)WIv7zt揫S5BF;hϒg^S-@g@!̻,\lqP bYKחۋuy 5lVPYzr'a"//d425FiN(Ԡx[AFm.9җۏ >_r #t[Ro#fEpi<;&OyTB[,/ d/Ky|vB"Y֚6^dӎɇs( .Ф4J6|̿n[H 3M*aKIFDoAs__? ts<dL+kc0:i%`OCS<*?,Hdii%R[mP49S%,F!aJp߁V ,8{&CD{fh`?n9BI4 GZ6Zqot||T݊GS r^IVC@phf+Ü_ؙܹF&qQjZI61 B+N²qxQsO57B_,ޥnqaor$L+vڰ@%-m&x0+M;o:˥ܴ1%6/;"xKXejՀGjvzjh3#@\\aMtqP\H^ JqTQ( Ȣ`;X+\$JKF20j;ǁѾzNǵC:8e$j&uaB+pA}@i#T3 M&j$8yҙoBQۄ/0e 8Z{8"?ڹN|}9:K"Q@!_0JX^s+58O!Yvst>ҀCSdm2,5RiܛN5{FIT`,ORLV-&J'1@ͅJ@*yhhl,A;>q᥇GѶqA;H@ATN80.vb.֙oB6Mx[+T%>m#D+?]Dֹ R PYlrszpc%ΕEV NBi`VFÄin͞G'Wcb]JvGf~HzPb"& ?ygLP>*YnJ 8r(nc,OסAs1>WX!V eo߮j&ޛGF㯣 jy;%ݓګ=Ewb~'J;-|hwE\ TK7j pK {TJ౔@c:x?VV, ʷCJsivx3g2B%o?~CtN*u\K ttHWUxٸWٿU/r\e]ey\K-sCC-@P$a%'!Bxn"$y5VTvNu^T; !Βw輨Y{!("/ "۬+A+}*t|| &5V  RQXJ[@Rj* S*PhlrJ;$h ;Q7! ߅,~XN~OyĤqƨ^TEz;J '[VnM}ϥOuMz.<-# "%?'OM1G"=npl~#KJK* *9;- _" WG6*NF;B/" 7Iy;ޅ4w^ܓC"iMia$#Liq" sKׁ:Żtv(N`8`/5BP/= $rkcGͭTBBAJ8^H\(5` -#l;cGׁ:aG晒 RE5j{Lb 9)3)3)3)|f eYC2-/z|y9hEg}]E.!5O|V'YLsyzu@>^9/2Q@< L%ʌbVhG (ODLoAT+=srWKΟ{ͧ]]Bk^+k})o"x@7ݜ$& @V ! ї)avL}TuV>k,fwfi3se2xbYxfxRwL.LPzWՇɢtۤ$RV OdOҶiy+F9Oq6pSr+.ӭ&&[OƑYR…!:}ħ#wkG|5cwp+#KCbefJσ.7yj;zo\ oF8H,#q3`'NRKXI ׾Wt] gyzp-j:8x{ruZ9p/Ze+OO:fE7#ldHNXԟ^PJpV쾔rGBXR"KN )@F)Bɻ4h5+$ٻw^~uA+"$V1& TN(bHqi`!*/Zگ`}7<B;V 5v4*}E36:QϩMϩ7lkڵvlvt$Hr#u+GM|IE/A:Th eZLQe* OL3eb2`Ġ%6h6DuI/!F@\`V BGkh"tJH@Br.0Fm#ƽ&iU2O&;±GPJslm o ?zGR~nӃuqt.p(Ȫ!P\`]Mcz77 ~k4f8y'1VmCK|_g͝_*J ż5IUcZ9P>mJ$]m@x+By1՛ ˤhxٵ?ߩp'ՌyqV~ N -J0 {oh*&sJT0UilHe>)x&%7ڣew;Z@zː)ڨKԼ1xk`Gkv?81k)irW2hih1+Y! kl= sŇ>6;v favP8Tu6ra!-X\.gZ3& yK<#d< {/xe׫u*:K1A8߳rLoqR+yNMؼ+@f36a%>c'~}/o't5{Ff:[ mM'j<@TLHn>7@륩|!(wj,ƧۇY;?ڙ;[H"AD4 `2Q68c݈ɵ>W'Xj`iO7?). ׌plFbTwE (CKBP ڭZTRJǜ(ix.@sX͙`YN' *gp'+0VD8TG)%ڔPK\k h&a%'!"$&6=m/}xw*AZ:x#RɜऴY F3*p thͬPH3OwZ+b 3DHF!TDLhxAAS&$wL ӞD3 JyfMU9l&5*lzH{H) (M!!.D@yAP4i pp{(rpFTum_ pK2|erlX.aPhe?3>!qs -jL 1\;JYCv}>Px!3C%Gtpg-h=Ģ1BZ8=®-Pz NŠBto&vz}035aKf佽4z _Uh^ K_JfKte٬ SIEos˛y{Q}8E}'9n!uոGw>}j=,%Ă[ ̴~{,|> WQQ:ڬT?5EF!~OWXI88o3Vr4qIjJÈAk!8mNP4ER9'6N9 Ź tNIP;9Mq\d=WkwlJlu6FPPc].x C2щMnѾV1q4Sُ?]:6RJRW:IB)"$#NJPQ ##6AD^}DbmL)،:f^ wjtYI\1_0-MP6,͍cPƵ+d2isp8b YFKUaR.K2HUېMVgJH0vrsWv8^oxE(O@QE&x[-XuskpF'/+s{ >ÄΐFg^=1i eusА߻d %&-,cŨ|2EOI?>oN[T%nYU,)js36;VA OgpjZpzCzZX{LLj()F"Q Ⓤ0E)u1%dϲ4 ]b)IX5TUf J} Dƌ6Z/b&Ic rƫ@ m](-x+G) ,elT)cV+dʞ*Q[7f /8uTYBy&mW7p@*XʃjMnK/sm1t:Jpl b2MS"G54ąUlsLNSeq +R4ig a5F8C(Z$x@ri+w%A`KU#TA4e,΋il:!oVSY2Y6vP4WL%Xwlp`&bg2h\|Oc{:h,ɾ^<`" P }+)f 14yjƐ4VH,@:6eb_7UUN5 OUD1 :i,byt 0,b9_]^c*ʦ[C':ʩll@Ȧ8[**dZR2AAﭲKhSQ3Dt0GQF4] \+ERl mo TE4VH=vW^en2ui)ո $!) Z`o(@H*+9JOQqBIX*+T-C^b-jIZ3W])4aȠtj Ң=X]dĬ*$ZŲԠ‰s4B u@=f8ö k9챮o{-*䄨G[-kQRq`<" 7 ^x1Hu@KyY=PzrdpҫhR0 Rn] L͇hyA3 vXm|Ƣ cڃchTP$bB*H2V***6VYܣy=)h^\@&6X,_Ix7VTlh ~tXtu+*@ɉ᫈;͊yn[Mw2<}XUpS6 ĝCLm]LՐZE}+pmpe,{%!z=} \Z}(T&ow@>rQV]R"hʀv w:&oC@GLrp >tЫP%l@Sƪڜ]1(w,pZnklT= N%P82GgƢN$0,4 ?x2ZF"vtDhÕ,cy9~}p^8,*VJH>K4К5''t9b,oOH,4LEרH RhWT*[fy/ztVIc-B@̤I@ 6A}$oUBeڋ!K֭hL6>Hcykp r6^v9m53 )f -骅 #I`d*mtnLv/sF]Ue Iwy.j hGɪas@o0uS>^jnhͻ _ '46~{x XRcF5G? -s&ʮA*BwPzB #RDUZ# RЃ1V[Z3[66I#+dO+lE]("\x>&P+fSoQ>u63n+F/T -J1<*Gf'UA`&~t}lem27hâHYG8csJ6Hi@jX]P/MM4byP`ҧD2)I1蟐e Ou3(Ws΋-|ڋ7FPٗ9H6R VxX@a}˙#-+Z0bԋ!W iDMs>FC{P=ligC1IXf" RLYY9-Ĕp:"tB%{,\NPkljJ\רvECީEF8Ø<`j?nz:w{qj6l7F hʀ~[P`YDs?w~=h/^&7T1j۽63=mc}YKPZVVs鉏#z_>`ZmM#ݱ˧8,8ESAyE x5r}k"~p?N :uok1[{2꿚<63՛/G}6iϕv;M&{s:D'+s#\_K2_-~᷁<=j#ŝ!|vLtu';_3e~Qzld|nU]E#kkQז>6i:i?i3NlnrGdU1 )Z^ƦTT$> <̜Bi}Gky)P"jcCNQLj;d\X2Z!~b}2~C$1(zce)((^2:^ ( 1ϵ~>]Y<}c4멙`*km]0֩s҅w)}>0хνy]xgb tDLF+A"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$bJ"$b~ELggj q.Lf|LfzL[ϕ!\6v.5ZuZmzU195̪+rcZn^}+!X{_rpuS+Vc`/bHՏc?a%DbR[.eS;] OT#^ARKI}Qy1A<4_0},ᅗ/l¦|_?bѦ}8lJ79s=r'cGm`ыLx-@^M$:o?i1m[@\e2t LQU*n6<*}3=~dS99ܷnfqI:=67{K:ީ͒Y•f܄%b*rG88}h{[cmQVHJDS-2-/V'*:JܢiE5Vzl+y䴂-E}瑒QPj;ۮ1UB{#Vd2Rh.?jt(ѧNDʛp4ޖO1yjb0AրTjТXtDާ,p搬r3MP[dFeX4weM jz@G0+ )4 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0;g위`# s>a XkԛR Ga ݅nvy}U>lѰ^TMYç|+P(~A/[ܴ>HIQyQGluyu>>^פccj>Rd) sȧN) xW q[MYyg bVAk5O(㵗ŗ<ǟ3zG.u6-/筆}e+/UJ,#ke_p\~GZϭP"O{|жwo>qc= C,9݈vAAqg(J(&x" :-aY|* /o } XL4!^?~,=z_"OW[_croeԇ^x{ܠCs|_R+?wc|Xر܁͋FY*!ەof6(Vy+x u9}q}îq::^^/W]ܸ?0pQr͑y p;"?%*p|8IMtQܼpض]4hiU;dҶ'Ū>87( Clb '* ԓ=U0z|ѣg)}M])Q;#$OUUNJo7A% g>hs&5ސpa31NAFV ک'1=H9:;0t41q+ c5aI-/,-O781"O=\t8:z%tQOs,1\~ PISY ?2#+,HK~09'nB?%FcpU )␔8|Ick{Xʾ©7Mοߠiʀv }f Y5]WxwO7S`qIwН`hu׹=:2O{^5skT^=,gi>^׭^͢Ch%vݮZ'=?6^yͼA3-WjP9o~ϱgsu6.m'Mׇ-:]<)Aכ6RHd4kQqu hIo e #NBUId5&.-uZ .Bs\l=/,&2]nk?YW\! :BܾoI@ S'oFhf*)IGK̼ڋp[ r0P{C3Z]˼0o^ɍj6|mC }CsQkKmkȾa vp}{WT!ۃu&(aAHDf[쐳".Ǝyl^n&{jڃUbf[5,ќH>F ֊#:BSDv S#4>6k?{`rE)tLɼ !NivA&+B#F{5I-]\[s # WT:_(T&l:$F#0rSd;\\]_0%cҷcuHn=@Sw?;XVY2 Rd IEm%fH;-}BPR5hMn6FΆ`l%_Lq)z8p"8Zu=:ZOVANīX |,Z I KtsůU.$-QpYN- "5ˁ6TRE9ϙF`IO&SQLm=q #/aTsgnNUN<῅Y`##1kzݫrmy\'Q?|?.?.n/sFv$.t[7 ƨaf7'`B> UZ|=Lz8f;^19eu2ɺQ/>uq#e!ϝ>w}?AdqNHgPN9^ѩ=?eg~wװ?|1tI_.?}4_`Q$ȭIm+ق'C5547iвYW ɸ)׌}8&t:navT _#> nr'8=jͲqGA;a ,u|EAR<= |3ʑz|P@%'c>ɵ}>GϻDN u I^ZOUpf"gLJkLm,KYbS[#iކC|;F.6C,2HiD *ð $gNXFA<&녓'js~DCFuFq|\GCc{bр7v3+GONkKx_ض9ک^WyS󅔯:=QbZtaZ/kRvBՓ3LLR=9SkWO*%n'F1JQT#9瞣D. 9h3Q#5"ʭh GDu'D݉oĈ?{'VhF,@P;9aCmHz\BW>jCaʸqJ" Ci,&1 M"e.I_d $x͠vz&::52M8ȚU9Vr>KwWBc=m # z "U9#t"P lke x4:fXs `'| L"i  RO|@W k9Ʉ:Mn,ZgҶNsd_礲<4*ɾkL+JR9Cb'E*$h1蔸U1H"%Dp|Z񵨿6䮹Y,j/73z/ uGV4K(s,T^r'9T&nB,h+$iA2eJ8WϢf {I# | 6@غDplT6aOH-^".rX39U\$#څh`f8)kj9VHtcmm¦dnui2w7ݜ ڴ=K{/Ҧb\~3.afdY7-%쬶3&+A>Ss os֒0YFgnyokATej(`?Z_%2f˘c˺:?0cLgXY DS͖T{5v*]GԳTjӰ ?Ȇm2.ti&[)wLRnJe|[ZQsnIY$awUhE@=j>fWPT{fњ>n p7~Xlά*nS@U{Mд3h6cު2*.#zهMq5`ųNPiFYg"@'-z)VnCS |ցPԨ=q68i74LFɜ$1n(4a؄{58^Ro0u܋FL|k77^}>m[$Wh6:K!? nCWڳ݅SIXi9*o|OY0)4F`4.4}ZHCӧ&4}Zky$ߓ8wV-E`4t2% ˜3B'd"΋$N$J 'rN sқ\[1)Ĩ5r:4haca09p z]p/51©4`ʣ\pZn~5"-@v%)UASz %v_{&y1OU$bOA"S-1#L:-޺XPPyvNS}:#b*Y N$  ['/櫲[mtL<}pGƲdA/dJxbX:`\&X - cY+E~z%hkZ-r3uȱQxgu&|堍GB\(ƔqrQ[$\m*m ͸^ܻ/p뙽,{*XP'S s͘H= `ST)-T7k~WuL'ш;`E|_> |8;koO.WyH; ަiVVnٲjʱEGBGR*&F^*z~TC,s<]yJ)\oEP{lΠDENIEILm:݂NX(A=>XX3<ɂ)1Zҵ[qm=I{lZ<jc)[-,@÷P@POMEsNn<o}n&0+!{K)D+dmTR>oi1^6I\/TpXȐI&vBX Fj!ӂc畧R*Ԩpgi!LγQp9c[e͵Nq ͻ|@~|]6h>=or`3Lt2z7>SJR-#`'X.<V+wN/`_sa܎\vr[j'~e3,qуB(PsLM[K.Jw] ;3G:~S{0NׁAg؍pJ*/ qm%ha?:/GLGF7QI;в9΋8 fj 8:ZgW:Rg!IPh`260Iq$ d5 ޵q$Bewq2R_ou pN/k}xDHٖϟEԅ%$Rj)r8]SUUu]xP'R)V*XV%8 Y(Xm1 F&sN:#~S4tsECwW7e}nkl 2& Y%>YX]!-q?yqCtkr#2h0.FJN7Χܔ"T>(LE**#-tk,=(n۽Ⱥ۶9m]F{%f[ -Μ}?I'vpH],& %9$p:#oKy!3Us!x+&%sA¢ .22K@&fԑg=VYXf12Cp,%g29qw&ݱ#3r֌ZN.NGUPib÷PN/2-=7M]F>qҬC`痮[huyu;to}=>'!78mVmͅPB>M%v"_;.hL Du.Zw3:w?zRI-yTxV:h9.Qٱ`u$JVDyZk+r֩^y-=v'L!5(> i?ŷ+]*$[*JZ~zG~Ico*W /A4%2:y^ ^{ l1 bE2w˸c< \3l =WJP0xdNe"3o'qi Z~4p'|buiwyZ/n/zU"car؃"M oRb/u 9ͩ=YZs'hO0Mػ2սHM?wbRĜak+.9zIoI#}UÈ8ref@Bp8j`YŲB/'wLN3rTu>U]u\1p.#e`EY,>co{Em=MO93͛㦚xǽ?="v O?>qa?O~v L00ѕ$IF(7oZv547ZZXghg θ)b1nkbmHʏο8L?Ӗ8O9p5"U0+D64?-Zq*#%œԪ+>BT& /'qξn_i]E;P~=$3= *rhZMT:!d< jD| IFs{:i6z!.,ɻG0-#Ik0deQ奰%sH*fUȲZy=Pg weC G[ :8ƒ&G;=2KO{,+ewg"<4i&H͙0 J5v{q  #dxfZX[&bp1.E$@nwL fx*E ]2F^%V^l \(E}NNgH'+VTvo^(kWkn\|qcqM=_=hd-ƬQ3 Qѻh$״rNd]I?* nٴ~{[UVtkռ׵6!GNҝoękd;ߑ;>zrxطo1V]jYQ|gZVrؕZgeQ px3-+EwH]BZ춫BtU]Au% Vq; ӀPW\ zUԆ`oQ]).ɂ]vK? fRz]D/:P?~zh24j暴Kzy=@wtŷ7F.?֡&r]QӅZ!]MJɪߠVʁPG>c0ׅG6a!BM-\W^ǣ2T[t@\sx)u91Ssdqa-+55Smv:_u!:|nVS*5AdD Axͥn[bgvBỲjݖtn])*Cw8(:+J;Ξz/\]*M5ޢ2 ,dְQ ht̘sUpd"ܷ8:p2g~Gru7d1 ?6  lQ}SyeO~ ,:΀$)rDx49h%+MAT7B'*%k6>j6>jc,ښZ5Qm|TGQm|60˧ɤmf7_~GB& -@JcCƦsɣb"43K6rʍ0$/% c %))*iB{Qee"ϼ;v "$Yd#Ǖ4r1qLHD CȹNG~ҌUS宀z4EУH9ZŴ{pS$Kf=7*)fE MMm( f2t\J|Hp6")HBGb3>Ɓܷ AӖ6"> !ra3J :+g'0nJADsRh֌aL!2E`̞%^ 4*@7R$pƙU:UƶHpU 'z-hk~3_X?x9ͨ R4frv:8]2d^dR㌂&Z+B'z6Ij?%Yאhdwᢛ|B ><2:q_Abp.ifL:9I(|a%{hݞ+n\rNO#\VG*R)}9aoBו m^w۴ȭ^W=cx<}ȫ+ߖݾQ:\?OUyG̯|T_bˠkpmPKl9O]FݹVHBJn]A~CS/b8:&|t|pN8u.hwFj僲0[^\=tiaz{);55y9m=VYXf12Cp,%g29qw&\\dr\޹;Ir!=9xnk Mtz.Ҕ\PdQe+UB/"D!Y fc❑sYZܳ &CR 4]2=. pN爙Ec.$,+m| V8i3@ \,>4?MsZsٽ@p'=-(t xxvtK*Ljdrgb aU4j:Q?'*vslXcbz<:y^K"I"\ O GWGEʫ;3xTtࣟUx}`˟ca؊\#lG-V7NP-EN@($68P9  / nli6Ng$|u-p<+tlZo\7n:CQW%)?-:zT9z:H>OkP0$wQMN=:^&%0lZS3cF;wCe˕GTq c\R\4`,1 ,;N-: ,AF(PC:[&P>ZGK!`1f4j|2 V/iZ+I3ڲR{e[GuS<` y:|_6k[i,SƴzD}I-#t;{vj/.ɟM`c0hˈ)$Yy#g}:$.uۋ /^>.Nb6 1YM Za=sTB0Ap[ur8è%i o`7[HmVcUIfi se'uR YZ&S+1Ѫ Vk%vVsoY›Tshl&)_H y{ߌro jSgp|ml7Y^N~@(Mtt[>xZht.8ր殺0 0x6XW+]%&nb_,$3IVrM=+%(V]:xuMr^n5^<M[Q=_;H{#HJ3wVLeMH&*f tP;s&V`9f;=Mՙ&P3:usz+(7?zPyX ?ՙ rrDNVNfx&%+elb~9ag45X)1g>ūf 3N1cOX*HMs);5 >ƮҐj?@Ѫ闣9DLzz˞L$^̫0Qӑ L}ǥՓk#o % T>JaUq2Beg0ʌIKn"Ɩ>^VE"T+[+zdc]$H-Ιo$WCZ%BB`=!I])ιDj[pZu"ՕZݓq?Z$;-^E 0<-y116x##D(VLSfK €)$^{%`)-kTxƌ < ʳp%NHRp+"H/Ɩ$X"zAf.%|Dm"6oȷM&m|_\"_DRi1s塭Vn+fۊmbv[1EZ!-IVny#uۊmbv[1Vn+fۊ٭mض9JT]5;K :Vs1@-uc,#J1/0RQpq! OM n@IjA) 3j2ڻݹ)Uvې݀m;rW~Nz>w/v{عr_f!G2  g\X f$/idpPC8 aK¨Q(Q0wOc*b^9,<ƙȳ=S#:qNFq3 Rq)ly[ü֯,b">Zt_,Y5l&?9yXeCX9ǝcފIQS J MH:|uJxͤE>byQ@@51 V s2 32wZI%HD0@zG L. *9C^^4VZ=skT#(Ria*A2de'aϡښ l8:EH5%͜^a c: R,rrI Micd  @ w8hv@%:e.Np(Nуg^ȫq0N]c@g7Y LW|2lŬ;.JinN_|1=+yMA0$t̰B:pb;]_fCsrǐ9aX!舀PDB J1uf4 W{{<:]Ϫ@0iC(XK ՘ 瑃B(T7;ɟ8 J$#_Woe{EӓdN/c'(/g-jbI'Іk''`TqZ~NT5wRb_γ-RþXiͣ];5|Fy2X  sVQ*X:TQl&a! 2$5P4?a\`6w*2BjT+#2"~h\QFĜ1F'4r3˷6P ; dmQZ[(DŶΞ;XNk,-r7" Z҉;/hz[@}a2@pPx\(a? $Ĝ4}S:(p5 (TIN  a[J, r@h@"АH1hYj񊐜)KO8KNPN9$~y"L0ϣR 2 H%̖ eV1'8ÁHT ƹDlc>" Eu*lnu"nNf}4YOҦ/Ÿ~!{g\#ɲoZKZGMV-,2~稾y~#aldO+Y [~$L ErIkse[\™-c;\[᫣3V+ka(]T^MJ"Q*i ?Ȇm2.Gb40n>I-V-J b;IY$Unq*׊X\ጎ~9&c{Eu2cgmN{=a:sp$M\><54i~ih0lcWʨZ&rnVt4UǓGwU]|93%݀#5'hwm[_$#\ f r0 $Q6P~2hT(IÉKäP oMpolg<T&FKGmi9a^i5`C}9%B"Wn1+rg 4i(B 3 /J Z(G T!i5Ȱ@MrR yl c^""rɹdD:jR>@E\ZmO<!@@eq<9Kde-&J'f E@B %F*9>T5dŝR}2mo BA=4m/…3I~b/9S,-%>1Vjtbpζ͡9wDZIHwB uD4ɠ&R:BV)97Rt`Ji-&ֲGݛ=rەIDnMm w5>d ^'x >3Wq͜ P,g&lRVL4Xbcf4++u ^ ,ZGp Xq6y>LGfL2U 5(9sR x8S7ދ7bF 6EUޮh-&-Ke) "4cv (-;}bV>J+@T>"@TO}b/H\eWkBKWZj]\e*iWJȯ F_:4\ʽѻ!ҼE;oJ\Mo&4zd{52K3rgX?~>Lh{C| .p pa%ε!46J˫^yGCN>N.:?Ţ[? & *PB1g6;zOK̩9-Z_nDS[jr0o`n';I1f6 {)pV8H! K 5.E10o0&R yx9ǣX "d 8 :[-%ި wLrz2LYC %my<f y-wP>*;( 'RHcԄ Cѫ$)fdrGs hBy0%OAht`p"+dFitP\fBcA+4vYZԣ"ືJ5"jcu$rJq#D'7Q h|qy'24";5Za:9$ m߮PO-vΝ1MZw,~|*oM.|fKRsw==۹$W2\.NRד= nX}7ndA)p,f->n={A^n[r;QvH)%. !?jAdqDigPydCW;{yO(N~O'oN2s??|5Οpr<\- jc~߈I7 ^]]s#Z59Ʉӯir5cAҘ톙-WJS6-]Rze[Y@vC܉ӯ )sw GTe* !}-xgrk=IW틕]4R`#)(' _V3*JQOuXm-\ w@8( 3DCFu8ZckIfy*ۛX&rk&%Vy)GK|L@}a2@pPx\(a?afm6O]YoI+Vއc6f5;~yʴ(&)bFV/I<$%/f%2"H(QXR 8@dS 09p\d8W˓]Bgyz9|s_~`GJb'i3m>Z9| pohbFHk6Tmr@h8d_0\w .}߶́)#' CT 7tL-=\L%mrkLN=Q9hzЯVG>:D<?&1Ȕjs_Q_@z?hu3' h 6y%' S#IӲa& 뙱1$r_jeLD&$S#FkuN_$}K[+ܓvpFa?iTNgrlGIEw}㝭cwvS_쪗 ZZnstdMT/"umeU·u MX j?hp>`s'ˮj,MyN,j&<_ysqKcq 6 \7is5E먺r2% ˜3B'2" ÁI-27]߲ fp.{0'Ʌva8nXRp0JJ7)EfT9Mz]{5+OĚbJQtty0#X!b-7FTЖ{DX6\8嵧:P(sTgGK`+'I x)X^aB0XxBVa}V(,V\sE?~Nd#q>&c6DL,IE)1AbX GT:}"R-C`fNڻocgu7P t;l9;w!Vbt:߆E|_w|ﯨ\Ա %gsH<&F &̆q*FRiEkk6bDmꪨqՆWs4ͦypZ?8Tעw}|12̂,k団䛝5{L F-eΊmƀ;9b-~W *jnu-_/+r2f˘c;uˡ+2ŞbmW^^VL+ګɪtQRasm&m yt ƹr$[I6m;b7,갉."XЫQGk(jͳðxÜcv 0u5,l&nóbQd0&lǼ͌9*e\o]Ffسkn4U )wBe(NZTլTAEq{;-8)}|yܥ$OX&dN716 pOzX||/"Sǽ(03s`r L 8Kb{#葔( ^VuK!nCjϩ|j8p>Tux{E:g`2M0kΪoFwy۳7 wnG=hWrA7df CG.Jµ8zl֋<[ߦApm^3TP%i7·o+(ѩ'hY4Mo0}QӫiJ2pwjSfduSb:Ӂع܌fGo GF= & }D>7%6A~td0zS* 1%ڬ54w:|w}.:+!6pMxTOkt;#b0˽厳K=g;Zy +]k;n=VlGp K%zk_q| B-ȴ`zGB\NJaʊyBj8-r ET6 9@>)p-dOKZPAI,VRʓN s͘H= `STkEQwզl|G QW5QK蚣 R\Loe4/)O(|NS}̶bʱO$8:!O_f (}88orM\WP /nkh[fiV;CA/N~S,5TAZ8 .Õ,E<F3(Q;Qnrcnt E(l | PRj,sn ̙CN)ĂI`hRh>6'V=$gkQ| 2utX6˶hl[s@"XˆoZIRAP6sσ׃>JWg -iJ."NBFd>1xkM ے~WYe X8UBj!$ED` cfZƴy婔Jq@4=gYQR^Gs0\okmZ]lzd - sSm\gz&`qTZ&:F=IT)%o PͩBs&($N;kd4Hc,qjk%N(^2 L<999Ax4LҊc ?7=j4qǍs)ְ?a13е}XWveP_"BILˆ1ፌ %֨|Q6i?)\p@Q'8+ 8)k7:Eu_ٶH@Hhqy~vuϕhc $7c#A33©GGPŝQZX&QMfI,+FΖ>I .Y͙cf7Lh.sNZb:4mto SxNѨ]vuV{vqu{o@3/\}|ӫ&6~=?͛~=7Lxr;[xlXlښǟ|".j?Mm珫 }s+˭d3̕.n|7"!#'_%s`[ tb0+J"1qg?k>o# IoG&Xk% cw;I1Y WqB:Ʉ2a=36p9G$NdIr.<ŞZDS*0KN#g}-IE:FԾϤ[xx ۺ>>Zy ɥ CfO0d&WS) *̏0d0+, ɵz^tA.LuAQSxDs fg/wοazh}D;.vaO~nA-3vm?t6HXx1;~SqBfrq*&S+^d*oy5Fر}F-ݽ (k0'|WY^۾lc~댯!_yc4;n5`xU)K.)P䄴-, 9mɥ'S?SԱkL%o FmKy2*̩L9zqT2lZqŕDϵUe+M/׻ [0SƍSHNsd1ieh-: goϢPT /WFIєT h)#Y"/d_ !nX`1EjIʶT/h/P$`NW]U]bF8'\>S@,cn^tawd<:.&aEB#h)o/E0w$cl7_%|OaS/Yoo"zN6eVLݻ)f[iBWϓɨ@G3 2k+>᨝4gs9liC&AIs:8҈yL\Ғ)TP:E`JeJȼsMUeqS[vjo&7VC?>>}1i_ƴ͹R:۴D5v\|I!98 ?RӤؠgp R -O Wg߹"qQW6כpgo*U(%/i/?ŷƗ.x_e>@yYyw0H&?bo*W AHfޚދw4Keޒ!Y|_ש.#C8 H&.nk.x71f;0( >a ȼWt r=5/:$: i4ȏˏx߻ _fЛ7١iRlu 9׏f.²ţwN\j4{L{ˋ]u>jvcl1^bn—A_έsI MM~XHF7nLWYh|,XV?Ø7ד5giUiԦgҳg c }:wc sK~{ƿ!f׃\^;~?{υ{߾ˏf`Rh$M"@=j追me[CxS m|لo3k >M3qg)%~ t=!;-mXꝟjNlњKE ݢTYxnR-BlČrY?7M>:Z6ێ8?lP}FLBIosS $cdi3р:HK.tRhes6nSM1Nnd] G 6NՃ̅m/{Qn ]ZzB◎4\5|3HnٵfiP i .yn+|],vCƻI[@Qty5-\YzMWҏco+iov'p9{u_ ~>iB#7()fv9TΩ-g$Sq$IE2ts`ݛg<ˈV D&O$($IJ(Btq&)vt?6GhZG앀+  ]WV$ԋ^ik9G]Z6`y-$kWTZ!tnNyP\Az>Q.Ju$H%ϩv!؞OB;Dԋ*־¢2MN[5)ڙ|]cMf``EGLsOa9Qt} cl%FѲ{+Vg=s,b@4: \t# BrYJf.s!e: >T?Z- $m ԫ'XؐEF P$+8τ-v&$\ګ2XOZGI 4X*3-BжSո!PdY)aѤZ:x!4ZR Q0ۖ%wT0*l<+f}RlE_ēo5rvnÝa5YawRUEǞ-1ZٜRyƹ%$zJN"L.g,&Í k2 %ckl뵔8㹲P, U' * gܕ:fs ~ yT~_|HVg8QXL1 )DƂQ(Rq|,1|ʣ&fֵa-9[ < 5b 20f0%`)BLY'dA }0a˜w1>IF$X.U" NAíD@PVm`aCCZiDY#X4иOQ[U=㶨[MuQ]ޑ<;? G=l]}I Q04ʒP)f^B*҉IJCEtQ0'Bmȉ2g~~PbQ Pݎ_k}u|5"}F_s> SI>^P"dJgYrL:J̅7"(I.1̅o-e}]W^z.ݿ9=Hups9Xы]zd m:[*}>I V߭]6 j6˔p+dYwm呁ՁA+DOvjc)K&vZRVf2;jۆ+]-D5GtRܒ tp;%ޡZ0"aiЊTь_ ~Č LlxŽrw |ztO3^3gYVx_llVS@Dn?\h{#)/#LhzՎEKs1YsZƳQiBZc5%-cJJR*d2%ō,G:@ZUFeB %¢ t8 99tF"90 ޲29[T"~FzDWp9uR:!H^9:Tw!(*uR٭N\6nPAN )#%тT<@K < ˑV$P& 1!9CC`B& C@5M*]= hα Ȃ.w9ZSe:"&B9@k"DӢٸ[N c)CSd.aZՠ#YLJaN"ͅ3^D+ alP\w ~jV2gj.(;oa6:\ ܁Gr:Y`Φo71u.} ܝz߬B]߬oo՛ foB#j7jnTzyZ L'IV1H\܉WDnLUV*T݉W(U3WWg# Ÿ**PىW)$c(i%,KS 5GBٌHjMBT;[2ZZ ՑȞ^=U%]jAڦȴIrQZƤO8tZ1*4NhUr<29;׏<]rJǵbl?WГ@jVs'$MZN\`c4ʊ`@2x!ic!Zk 9xpL;%'KAϹ>R^Cpzp2Hz|_MJz/ ?z/Θb>|a$lջ=Ѿ催H׺M"_v̮V/'Ft3vQN k4vwW%ITV<~Aw^sc6 QZ9'a,P LM9O%VXFEM1{L»ʨ9YQHg+So/߹r]V:vjض=a:cL?/#*FOh?sp޸ʮ&.FfvuF #eear%RɈO p9B "t5j3#yx1RH8xAgaZ-Oܾ\j#=ҒxjeNU#!@-[,8 - TTҋ{w2k:lB Ȫ>%4 sI&!CtR+eBx,K1LhRh+٪Ä~&zdsz/F9 ڧ84bҾ+Ӣ\ϧO-u~v7ߖ?krO9eDZs o_[m@vnޡQS#O%BfOWgfPݒU:Xl GNaCg-[/&լk!sO5+TnS~T3#-Z_7-]$*KL<@X.@]>O%;T2Ղ;f{gױơKm7ɢ7jBS(") ^{UuKARR2sC7̦=0cp[4F3CJB5rL%[提F6OOƃzӪ-Lww?YݝZK LccљbβИَ+||1Z{I[6*ԑ+~[w ;3dO%e"&ֳu5&\i-hyc~~(`v7iG}XGĽhM2c董XOϚ5Wzhݠ2vȩ9O{dvXmxn}(4D`jBJq!.jۨṅXX |0\exW37~ Nˈ+-mղ[?+닔  d_y.EF'k  8P:\1 &W㣯1 |P.LoZCI4%` w2YN kFS1Vl- F{\ݎípT#f':㏇u%/Q\.7XAca0YR )x6b).J+4~LZ)|%tK6#=\Z GMF[eѨ]|0 IAkm0W*ͱhˍ-x>[*'vw2u_lvlu|!ctmE\$TkI><]?2ِz G̓Z PkAy)吃:8gV5I- a 3BmH2#lGdbr<iͅ+z "[DtQk7d swᓭ}.$n!H&//#5L+w7I&kndP 589h,ƓFvg1u(s{ZJ*cbI#uR49)*cL˨jkjtk($gVBe].<.ܪ2R6@'iy2~߆h*=ȀEN\фAWFf6U9L65|`E;`He HV a "ihP<Ϡ%cv= _21k>Z' ٬4QEnI%ŴH&pA( 3jXˆ@/ivLIh ;PT,vy Vm_VDu.)(АYUhl5s y%%#FT$fD'@_ɿ'q*Ӆ!oo-\Ի@P5<|_ٓCMPEx5Bz5+V%?^f҇[ӥ:q^D@\̐`2KVWqHUƹLmTuѾҊ}7fa:Wߞ ^ w+ýޟ_\fCrY Ӑ 1NF1q02\( (%i$  Ȯg[:u8,|>#L=NQi[|}myW\ry>r߽d>#`y:{ m V/,^&ajaCW SLiQ/2aww8m Wdt 4w9ix^sw?.3fb羬HŞ!~>2 芹=r#?&FĞ#d]ҷZf䋭'10mf,%4-y6]Mbީ 6u!:kWr+mǶfRBlbC `9D(A`;bNپWxKP'ω9,1!d\NWמgR.l:{]u-vjr(ɕe9f? $)%B8֠Df 7)˖&2TN:/*_-W #]oIR>Ф8ڥI =zLN ܤ]ЀȅQFT^4z)dMӢkC[% 'ԇOFX}VgaZ%4d˥1h{ ;u[!*(sR$@x-\Wdi+c]nWҪt u\R|IAV)e=BgYF e:g&!hdJ k9I؟>@#6:zXT2Z,s,fw,|N4(Yݻ%\;;ףKwOH8384AuH18g{ח];k~%-‘J@ uT5٠&R:JV)0:p4j(x9vG'&&TRi9emWMwښ>D?7E9P?Y}t @sd+x`T:x ,10n[* ee`Q{geLYKIVg>1M&'Ǵ\% UNEN?>Oyn]|lާXVpo+zD唲m"] DAgmv+8B;2@%8H02ixXe`#U'Ɏ=E:9dEN`Z y͍ADinl0(h%' QQE %KE+=1q vZPoG!QSH@9TUsp+AָmwW!Nza.4=:::*})kY_v@$!0Blɸv7!YK1KxM ΎNw?^}+m#/ЛTbTZFŁE@ wIMs-q6TǦmD 8Х .t},mFZS*>%͟^v )r;߇%gm-8SR-nN }>.B%@}&{ ,\oS}Upu W>SUKp1cf;#<&^jIۈJ,0U-KI 50Ryo@b<L\ 0OiU-jp4$Vd|dFKYQzȽ- ]>!rx Nd9b azםk9}2W3o%g}* Q>@clsю$#x=׬zC׮!վwK&>pby{G^-[s,cL y8ω?.t__eL*7z A7p63' ޼rpHA[uB ^`f "* 'A`#OgƉOo.[` Q܃`p;݄QvY .UuڲoB߫_߿z|~/o#:;o򹆭_> /X)ɹȫ6HĶ^`C"9Lq#%vO̺GCO(@x*++R֝"G I:ɴ :!')zp"Ġ ~΁ӝ!4.U=+rPD,@_ E'6璥1+Gޒ] F$Eqt摔ɵ|ݓ6V߰ښִMB_7.U-F =?^tS>Rlln ϸ(QN\ @T<=nǪ-~Z rj5*7 ܻ,,^bsߨRmj wH{Y?Z3]U!*CNh9S4!>ȣiQs6RCl YBm;UYn}t:ggQK"C @D@ ,\B&:]Be&OMTޅ٪jyugv337J˽l{_SeI\aWL:;/4;s۳D.5V]C;ۋ#+%E2۾Ԡ" S!p.nŕVS.\Y=\Cd7г8* w@R ^LpߺӸϧ|_|oұpVVйuFT_ۛo}^پǗ J,SL3!r_׋(rǯ.8lSyr}],xtg4<Tѝ_o يX)gqߺ^]:Aadxo=ͤdLBnV?յH?]R|=L6ٽ׼"vVF:yɺYUrKfZfLh$O,}pK R'KzA;C??~ kLkntxOSO_>'̧wӇ_ޢ B.7_FrGS'L͛L͍fj[7|yM\3#Hȭ�2Lʇw*dLj|`]A(;q\6(L[vժ*셛X>fGL9#r@%z~M쉫ΰ~V?ZDO3MxJRPND~[d@Ȃ+E=AcTp%|R HthkÄ|?EB Ps{S(JY Qǜ )M{3cH4.( Hbm:8uJx|kjhhL,f>W|葷vⓤ3./M7(׆.-t]Z/I}?K3;K1~e"6% -D0A!4XHRXT^#+TwHʹk%eD5!sdR"ɘƒL2W[PMU 6qRA5ۤYX"xN}X(Z:OFk"k<믝Hkdbo^Pcc+Mehuv cH<(E͂BULy:D(#*@$ݨAؒ#BH6Lϒ"8hkqy^kbBþƖDjdE]>JpcY(~x;t*1E_f,x&UpT#6<)Lh״{|\4ʕÚ(FFZP>9BtWr'ka(}r)=1ҡ01qʐ@E_uLֳ UQ*KȤ挍-g((\^(慢{wgqMͪfyPaрz c+ن|'ې?32J%ur͐MLXVmnr6sl?9Tvc\[5̵U˵[JRi4+s(B4$()t $# v #cYZkilEHinDyԂLH#@ &&4w|MXZJ'5(=k:p|`J&Q)ML $N6JYŜ` j"QS.}_gOOkLTe'֧n~nC[%m>}$W=[طٻ9=[)/k馥E*(4$id90KmԚ3ę!x=b,9A]`4TG2gjBnR" ۵tǭ~ZEZ8bc21ԯ ּTKxe[\A-ay<\Ҭ[ #VK&^0MTgyQ0mWb'MzHm6wOXdCҶ iz<#FQg\ܒ tv(9§&PNZuH|ÿ|v'-_ZI4 (,!*m7Fs`(RhP[@'`L8撐&PY' 1Eu-ټY:&ǿxvkWΧ5G8lmٮ,=Y@;;DX~1g4!_" Lʛ;v.4Q]N~nytG:Bhv9B$/m+6nji-vJ.i&{`Ks iͽyͽhhr:M/AlR(s"s/ ≇@H;R<#E)ڀ҃1 0< $nxh R$6 S2 '2z %>=0)&3#s;Rm<-H&ΖGӬYqb ySUuswnz o_u{B#+ґ+WLv|bw$]4#)I]EoGƱ]vީyQ YK%χr*\fZb#i]ZTT+:ĜȭHz'N;) rLC3g2{|;b(Sݞ]U4Kx|zx<CV9(,{*`A Y{ Iz͊ "1BkmQA(˕1̥PG! BYg=פc n>`9eܚ8[bqgMN$@pkuho龿5^hs3nײRw^6l!Ge)DrgVT2z44i'BR2X, "69$\_0#亅f.)YhF'n9Hc,Nf h5\ɡI9+HN= c6{[źǩe=hibA iL i-lVqhMLafsmM,$D%"Y))V3ApE4IYkdD}Bjw!=a^Ӟ,s9,*?*ƩPY@Q4*R[r̯*.*Ej:r!KSM%fhPgZJq_v:ѥb|_eY,|Z9m #ɋԛJ]'YZ ^ ^{ ;Y4cr\.+zh!'npC]Nnp8#]X7.J{ub+Wݿ'V^LΧx Zsyx^.9zI|piRRԒV-IuKMf{Y,oq4U GbG@m/5T7VgljJ]u\2lNCfz+j㔝ި>w+tMueaW8Û?Շݯ^yq_x+q#0IN6,¿$bONӿMfM +iZoӮm Y>[U1-Z@B~0~,LuS&Z^}e~Yǽ02s/\JXČzTPAxo )'q'˹m_VpQB 0\ 3T`Z)uWjg$=aH|SQ-QkP) b,3u^9Ѻ죊2e׉f&PCJeڜxT*;\;TvŮأ4?Qnt=KpK9>ܞ"1O,ՖH fW]f2I0f,fq2, 9 pH7/azyb2YN׹ryo*E{>A'{,eB`]TV٠*4q]vd(,_nx;8MjHv\dSϕqŮ ރo;fOUSc&eภM+CuނUrh­D:w{][58IX5 =Ug[5"6Y!%֊gGe]*4he\wqĽByqKk;N83 YC3>-߲S 7 @9oiTJԷ0D(ORŕ: A XH#2 E8:8 }@( (P\"9-qڱ L) K,YQR !;!YĦGfhī\G養>Pt%uMQ2lMPpD?ίԟ[ Sf3p  #?;6XTfA21^^eTK+( =E:8izR3%pZȏH]BqeǢ*RrөP] *kv4ꪈkı"=xuUթP]I&Q+2s4ꪈ+䱨">,EFqԕ0j|*~[uO2F^/*?<}d3%ΨWw]&uш93"zqpn5%eӏ7<F8dE\ŏEMi"%NMjZ)XN7f^{oяODi>NpsJ)UhcǣAE=VLVH=f*/eR/7:)]B}(mN5a+2^T,_ŝ٘OO3;F-R 1D +$s}ܛdtQv[Xu4mWQT5zn[?nkP\()Z}, 5Uw>Q]YFY,)Uae2)/]3* H&>1A8()AUh-ڗ/xG*@@S#9DǬ.ʬ dmI/S@;fUeEj>l&\u^nPҮƘ|ϖ@s]旘Knx>a.qɨYi컏x9Nc/xK9¤/'84~cR.):͗dfމ4c^t^i6WwNXX%o4PH9OFrWQSE 2d#(LLy#` sy<}w0 +Py&Xx>#@;V?\iAM~7 /"}tz]Է`VQ?Ou(Z9ۆ֥vfE 'jr9Gd(I)H;C:K-pЎvF,>sG)9g⹗e3X|~b (= YsLP@6ƠYh/Eh3.-IɥUZod/$LcȑV66Gԓc[g\n>x|X*LCa]ȫ9x4~ɕ[T8('rDoLSY6.}գIU'ZNm-ZƅJ2@\W]jC.&i3o&X)M7T:/!v߯ͪB,I$]7wj^\X٢楒h>o3moin8:4hwCvkCͧ&jZ9H߾\\tL$m0L$SP˖[Jܓ+QQi'̋q͒W JГ?m㴴xH W,˱OǕ}~9cYqhB1YFA3.((eI/<:$2h[MF MV4`#,j rnxi-^>i`[uWt{].^nKץut{]THѥ묚.^nKץu~ 2ݥtt{].^nKץ&^jWg ,J\?:V"T%e#^IN Z.14A:{&!CGʷCH=X90ViRVx4Y;qWĉ{*`A ɲ^czH,0/օ2hhj\\J {I- wa*阸X@lm&Ζy~@nש~}Enս͛*6{]M p+6W]E0>6՛@7ΫHQYAe!;4s9&ԭ  ] RKp5)  O! GR$肖[UIU%}r)ڼEw52N/=*~b)/# qŧ[*(Uey>gvI~EG7PML_pahzTq$4SSp.St`:C~x/s V$uogpo~Լq\*S.YU]r˼eyv [ 矊/SxkYbN.#H6ѫo.QQT)nyϟ B41UjiX]mOjR|btۚxZ)t/|N pIPggkok[n4t6 +y x"ܚtsvU%&(-Xk 6!$HP&A^ JP\Xd%pJ^kfE4 nkF{'^E?v ^A) Z*M-T=꥖ Vtozer5ۗ^jH-2o{\ R{i.*{/[6ͣמZ(MnT9v 跽Xa2T5=_yQnq-4ߗt"-$yZy{P%!q" y a*WNsU+H^)rRD4TS5DHG4Ϳh>q uJ;UvF"O'U'Tqr./;g/Q.×ԫ^OzW:EjEu^*PFITF3`9coO\"hudZNQjBH-,* '\Fd 6X[J؄ 52vFȸJ vB b³j)^\jgiJ;^& #XlD$k4]RD2)yD/ʛךnk!60ceRy NcJmc"^鮋vEƣpù;ڲcԖ=j2gI?Eڄ(6AIHQ`, m$B{(N0 wZ1eH>GX"Et#%1SģSag=̀XY_^L_*1w0j2}&Ph)\@'y}_"o/sFԓTjf=N ";Ҷ]`\fKr[z%#CrTio] IuDh~.K%QY0S³ Ш!%5ǯFGy{tF֙ ԤsxZ8{?4 osŜ2D׉Y=Ւ긙ēy[S醲Dtasf:4 ^ϻ=@iq؜vq+.9c|"p%w`ܨpl g pur^I aϖY-1'wKFgW؎ly;Bνzݧo)pZw4~Z\xY!^7`R2iL !R+Ka=~s]NM_tٙE翧Q g6Yk2FssdGWR ( )Bs)H"DI( @%J6P;5F=I(@' sUh@I 1jMh-l vlvF U_-dl~`m0Q`}$V(\-{{ +nBUk{:y"JS7j u$”<`WPv)<1Δ\9u_XϷ[ڇ: 6J4D54rg2x R)1Zξkvmoo+:ލvZ?r-d[Y}W𽩷ӛ3Wv߭3&s,9l< O@xyX TڨJ JkWvmgrvZ} ,:~WaK=?AHRh9) 47.h(Xj#jk%NR^rB<]Fg<.Y߀OYj_[zAv?Se޳ܒg5lcAKy@)2t:p,"XK"J-ZDtYX<[,׾ xOAI" V2DoS! h΍$FZ)y,̓x:莪†>T}m.5g~B j~#d4'(Km"!8n@AX\/' R/&/˓1 =S5=+$P7pl_*S+ԮUR^!\1Iю#b?\eja筫Lͤ^ \q)ċӺpK2WZw2Fp [Y`W\2z+Rk+C T4\6ѥFh +w8^ H#‹$Uwd<\!>3$ H0?0u #QR{1_$oGբ4d:RaYG5# h< tϢ\6/a !Ar67'f8E OnCB*N 6QN(*.gd1σ InMjx0_%wL,8 QJ5"V!o!œN2sW5462)x"2X) toLa$א13|ML"J &2Z \!(CWZJ5•=+$Xp7pf_*SR) •F{U~ D~ѹi<(ןף9[h\i?Jjê_0~震 aSt.V%L5sݬ|ogd0SKN>סVyo-uyJG- @wA{Z2l-ļzReĖ ʗp_Fpx"#_e #&:RWZsG!6#Y/v_ !ɮ 8}@KcTHʶTϐ)qHJl^dLsUUwWNi}pGx͝QgL6D=R! }ի~SQj,ΛWOНoѣ\0Y,ybV=eBESWgST"<v")k U/ ^xl(e2 WtX4IuT0Z&KQA2@(: k+Q |w XkYzA("D@B1MJzjB!UZDR3VY2i }}B{0%OAht}fPD2VȐ-Q#e-+M|4c+VzAV #$ ՑD;)M.!ht.ݕGW>ƾ1J8 #I6ZTMO!AkT'圍;E*HO-5g*,>:bMQKC @D@ ,E]]-Iey ߻)j:vHdFLCZ q?K**^o|OЛTV |%r&!M)=ӿTS"~ $38b3j0ɡpή8<x .m57(NJ߄w@R f74" ԋp8Y~?42ۏOpۻv[_ۛٿZ^>\]/TXfBdγU+rxOJ6?nO|D;,.{K4mkggTks]}㻛酷l0V;a ]?|lƖۻ]X$ n&5[~AH[OJ$ꞮںaXdyJa0`Ň@_'k^N+#{d[m*%w:ά=0ӑܱNCozl36>CYPs_v/?t ~ww2s^zL "Eu'$A]O蚗u͍ئkUM+@}iH9~`.V tA?nٽ^](5Sf3z{qj~>#LQKRin^ 13ۑv|Pox 緭 -UX/Ghq%Z3{ARPND~[dȂ+E=AcTp%|R ]tՆͫjW@7u⁨uIiܫtCqG$W@shSdSg ok.4K de`QmoK .㪎(FeīibYM)M84W^>Ƃ$+CBs*E X;o|Y`IkNhƑQ ѨplP= TQGx!(4$i|aڨ5G%$gJ:S) ΒVg3C8VΓc8oTGhj$pq_&~>)W d $e Qm >)ktT[C|Ew=ӷWO&V:cpK(vp |EgGD=Lgr`I] J\1YҡT3W_yY3+. ' 9HLb5oV>& AMԃXS4o,E%QEFe4V6&RRƹt[$)}=d*@=ƘgZioRg3oN =1i־x5ƙy5E8\Q@ٯgG* vHp}P>]iLTZ\hLPZCӘd)hL^"LZ&aot{^?*e&pŧ KTJD"I" ³}P4QLрӢhx\]q98$ഷ$DƮQ)FQl-9nJDx%x&l G* ԀU^3.> 'dSV2QT8r-&ի|2|~>:` >n)|ۣCh 3Tzks/i/fdtqkKt?[w[b<ڔkԕ*!ȁ$,u<\ŸU,X )4BKQׄH$cWA&xUwVЀŝ$rAI *L9- KC$Ec#dq!DF blƝCA~=QvCm3E_9(HV8Qg.@)CJ"$ٹ(tM4RD4a$$D(%z zA @j $g0MKCD5:9,2#F)eӟzuS/AK]8%HN22H\3 Z r#:;!A!# ^ 5?z1[X1(׫x1拴>B!0AˬQG&./h#Ԕgqk;۾`ee_ńߞO`c˼O")},~,A|j5_Ԇt7tl+ Ց , UlW MMC)?x~ 3.;G0ύo{oMGf-] )gX\ꙏ`pJ&fZ>}Dy%,w)nۖ"׎$)&5tdHWO.8 '8YP1i@|HT-b\*A`"` ; LO*Ed24@Ih%ƣfAĄ;٬aTD W]Cp}{*ɓH-𒤖sw趎[6X?srRyO>rzoÓp<-/?E&AiV蔭a"ofk: q擄}r9@c 8ehz :&YtMJi*(%4R-c1qȗ[Rb ͌BQ%6K-a͚kv&oTr3;؎sv^jbdC>jU K4V6+l)yY O!.2½geɕJx%Pup J ڔ)%f퇃\.6;jUa:!u3% Pm͍46JAJD\I1D"DqEaJ9bDdHAD A10R!p\iu5&Isa1q6:xfx*xEԅ-,bgx mQmRgv!8GDE6`0d I)AS3PB0|4ssp"MSf9<-4myzoH? wĜ2ni#{tk:i#=wBg(N:T5Mˍo E)ok/@/OTdžD  υe"@T.{7Yydc&{3Fޚ镦_{4/*pAWPdU4}y `ȤТlG\_].zqy>U}5'< sd5ITZn>* | {1$Is4I}b¦h|Fs-EHF!tZ5"GdUix6ڊ=dxZ"Geؘ=c3Xljs&'sʑ]U#`'3M7IlC-$ʒ0`c)OQK "d ‚.ض-0Hbpt!DI,F$*(JYt:Pܛ gߋ9ѧ"TVQB^A'MQjPH!3h& 'kXr(@)cL)HcH"rH]%Š#a763-f4C1߬ͯnf7MW-܂jk;7f_vZaTg~Eϥ,VI GJ@ N* ;Bd09cdm뤭nC,GHWLqXT%kRBּN͆>g ݭho$ uՔ_ %/YkC=ALkOYXGrvCH rk$"MʺK֨.dKq=|>(o42{񈞒rwuMc4S}>zyl1UiV.QRoKAU&ɐP;@uS,\BmAYe$޳6'M!1p%IykOYr4F͆ck;k_ȝu\GroStu˻ybW^ykYi1^lr!X- ,QSTM-ɉM0o 55|S\dK"s I[k?px5fu,K>XxG{ O@h  x)=KE:96'YqԃCqxᾧ,om7a ֝?/l +h=E&}i(6Ld1X_{Y]Q&I(!UY'1CZ9$/D*udT s/Hm1 k%K*r!FZ)%C)ʏˀG0N"űa^Y#E"~^V Fg_KSMV?g>?C/j|֥r1ǯM%8ݼ_'qM~ܟU8đ ”i ] CƕI`C:bqB]i-5hMtzpJdzym7o'g+$A91twhGsGZ{.'}UM`wfW#:ucv&'f &X{3tTOfsr F磟&3oy-'wH?!ڞuݨۋL~Xwz1>ON' W%6xu{Vyq dBFjǖ_/6}5yTq,|__7jĆF;.℗X_~y'{og`BXO"{X|uE.]O=]߯NS\nSW応^F~jfdGyf1?Q{8ob:fYP@xGi0݋ܵ/6Gha+Fh¯Ey9G|hgRq)"GEL,ކ=^N8e炥sZX)ɡN6j 2Ē͆JD""NAk&UʎeǷ,elQ9eDymҝ L>i?K)(hi嗏-SFLap%bP`siӋI]=.y^ئ5Ñ=i`do##S3Q.Ť:I%"^$EՎ^d]xE&J V)[R%ǂRX%(lhQ lr&^jCBHL\Hʼ؁Pg ̘䅫k7(_JS 7 s~2U, zFh\/q] +[5APW:M*Jw8=8 w@( (Pb 2j5Er%F04JZ+'Z8OV{ktS(JT!KFU`#06)K%PpW9M46'dMeKc(l8ۡ\/{\#z;I.oNʕ)sV3{-E$nYTZPebz'8/~Ѐ=~p[AuxFy۽|Żwxc_otw#uS9=T>gj=|H^ywJ%E afLeD[lH֯5m^QXʍ(<,9 C2Abd"a%t&YٌdZL^L82w~J 5fF+RLWN@ Phd` Jo%¾ $)J6&vք(U&  \DK?R PR(|z"h=$P"|I%Pf BD̒ X++Q'_rCe@w#'W&y_"o&׮d]Pa"`]z; {@ kqbcA}b? )3o>^-=Xkq&)g:(R"w[!28O9P"|Yz0yv1B޾wn^'(jY  eщQujl> ZPf} K'ݷ nKx~5ShP &|_?4m8 V'yo^sati>6'/)7> l\0oGT>? 傺f/jKmtƭT{TFg3;/li}Yz}7PBiw]";tQHiA!2Cтz5/^;,3b>e8zs?"E2dNcJ IUDKtd|{Nqؼq&E;0o)ɇ%\=O!Qe(hHeбa|%pf)cJa7|H7 KBQ-g+OBdxL+m54y/G gYJBq!>bfQĊĊO(;\^O-5,659_mr~mKK5Br>^,5)__-jͬXtoʥ,yu2Рx_O%:ssII @.Z*V=h%$#e(DOIJfl66f+㱺4օׅϪ #65[Sޫ'ɟƎZ|L4U,B,Y9MdprK!VVdcȺ m5] OCbVBjS;$gx0!枀P d+HE֜0]cQe\n:]cz#M+^XCF Bֈ#ބ(Mq*)3^ `fT|VVd|Dɳ$)b31pZe5QM>péQ`¬P4bXkDk^#nZxe 3;%eJ!TdPg끀MV,Zc ^F E1Ye:{J3l/"r,yx4X5UL@|Q1^?^^ubIᰆz?$JITD1NDf<#(b%CQ')h/\1Dp}Z6C^@1D,`TXrOX4F)rZ=%. : , ~§9c1"y^Tss*݁-VwӔ\B>-*f$E[o= @f0fFBNm'9_xQ<\EŝqL RLY@W'<)|<愗kEeծn7{wH>vq몝/bC)_B@mW:A) #4RʳIedx>"`) t)!S]zJ앥Jq rgTr4d7|N6 LaaflXfءBGa^k9;͋yZK"i&"pdb /T?>&o~O>KÏ~͓NCdrQ|g\ߟ{4jŖaziTj#8A(S#eyEʰjE 'oklucAu?퇈[C4ǖ] "ϖ"8>dí Sԝ7O~ҊLӢ:ߙ\Zɏ95*mQPnYaX G3RF?P/ѭs8\8f)thy!< N(r2R'P:LL,Ih:u]X[%)C#*0s郕Z)CkBX;"D(Z<*)tBX/t)ǽʡ-{3\\<̰` yyT,mm `,ӗL׆﷿% P|,ک2]?x-cFF]⮠[ 2^f'N-[8 Y:8"UHdB8)-!(#%F 9 Ym#/+UN2ȽN´`yhnBPyi`#$L3VGGhf焚yiSDvxx@ጳ"P, k"Ho"uxfRTkƤWܢP%Oڮ9%A x4NҒ#Ǔqܮ,,%PvzlKYPG>z*@sfspRNI +WNin%G%38C!.@%Dy &g[yJk }MWo|BU?6욮]x۫p8'r%iuL]ye3ZŒ ߧV'Dɭ˺uY6Kt5YݮZ{=ilB>] x~8Hx6ڵ0lXcMןH5IM/.S)ZYnmB9H*u< PsnE*AM~aloS ^Ti .Gu}}9I2}t!JQ^Qƙ Qz'S,X5$P'eY8,sr.G$J")!gS6sgZiSYZ#g}_I]T+̷^3^>DOb ؛٧Hro dr%ݗڅM.T]r& .}u%r-7*kؾ+S*SUFuй8-@E/SqێiXl' hB ^(<)~M1NWsl#KJK& -?1E5E}wx@쑖F"{3fo@%Ru-\W Nד^Rˤz~y44C\PQX$?p_ߛ0D|Hik -㛁bm֜M׼q[4J٨ͼcկ߽{Qk.E{m{Zm3ϊf"Ȯﶙ1n춚*l,p2Ϛ.I U v]]e*EFueQr,CVJG7<;ZZnhquI!4|h2&%5KEp* -r)xɮ9$8_GB_&WLkɓ0-NP4*єjhZ}T d"{52;.)wggAf-u\L8QGy?vM^ĢgV3D5o =ٻMVLNtTFV:QUr iFHX"c8n_c1LEVG߃T*_zq{WȎAut݇oN/Ny2sz'?|;_p& WKIPO&'Pg {$/547ЪY7 ɸ)/&3L5n?r+So]o*WgN~|׬b> =ӯ (旳|1颢w/CB b3=TPUje>*p+h MHʊ }W!8h&=ZE\)Tp%\RΊUXIm꼲?w hMNyCNKh5!y٪}O4刌j6xu@q(O\vX̰g88 t&aZ)ђ&h(IK mEDtE:M6ufy7vޢ]hh'-4%=H޲3Х͎roғK/1~,K^'K5|R>m&AYtQY)Oٛ4ῥ{?I;}9(U EC( y'P(^ !"fL2M:qSJrI&%T^fF+k EH@jt̜41nXQ/G{=߻z >]p~RqegwLkZe=+ f*c +49 fndE.32ߜɿaۓ7&\Ol,W X@MJ?H+*Hw:BTd y"#rT휤T=bւevFDBYtb'+0D@mP*D_Lv/-nG?-4'i1_Հ+_j?joWW懼9>e߫_?ӻӞjKoՇ7G$f~GHǤmCm-5/_*5>"R;ґRR)PxI:l[x)JZ3RJ'F_P)k)[y'}wH_J \_JLQ3yʬ<'iխ[Sw Cgr>or(QSccURwEd(1xأ.>=S_v>.f$åhR%䒻; d"^(Wl4IaUB3w}6gn v=ft;2`2RSΤAQ9_x2|V%W\Gwuc}~R`/I٣ݼf^H UwUSdJJ%FQcZTT =KGBNX VZ!̜ȸ cc!XxR,XTmwY\DOD|2]L~Ln_;h-Bt!ZԹr#1)sd62u)ʗB^|҅nu~ I Ҫ¦,D lO'>ru1VGx7M[PPY8mcԶ#j vWW0O jK4Q"`Cbky5xXu"Y=FfQ&cMʂFZ&jXM$ATgBcٍwy>BZαځpE@ƨT|K zXkR{G4u;iTR,3GXC"pL+/Zʺ-N\]AwV7O9O̸3~9 r;hE ssl59ݝ;﫡2+X5EeK29YQTU;ŬCigJ$fb-2)=(H$۔(()gDa$ЄlLUsYCv7Cܙ@QLokEZrC|_²TUϊxhd{uCֲJo+,1DRI+Ϙ0'x7O$# IdɐdR48~k#z0`d215yYx43dH##%񄒠kDrZ,H*@&JH)2)K|e?Cn } [߇"X"$@RUĪk 5B*stl;ej1)ѐXId[L%K0*;C&%03)`1:161v.̏c ]8}lX w,Jq@.6JA+5f_i=OWZ̠=XlW?A@_1´6ӓS`vV{`ª$1|T]Bv59ʉyF}ӝwH[9abȸ8i  a4&} HěrkrTgjZœv/V5{/=GJ Ͽ4 YnI4pc5p*XBbZ,<%cEr9hI<I hM-쥴 |YƂ SʐՁU!Ѻ L3sywwҷf9v^;@ebwW agZc@^ThY_ƥ|Eg0aiBAS\x6ubpw.mMM[:T no bY"fes}3m < ?-s-kIWeBuk;蟏@s)Q|T%RF)-YdUHZ EhU !|$hL[iRm 6:($m_&mf~{$ɡuՌM-8wܭBH1C1`\'I7zʏfPhEtqoU iP-8_Do;R|t(燊yܑ6g]|O8bq>v ا-]&lMBO#:RK^bQH/Us%\RVK/U+~A%K+bચ+UV^)ZiW \imV rU5W]eVZ # קMlU5UapUz_[~km{PWXW][I`|}knNÛZ\ B_.ݟWweOs/gu$#>KFJ=}}^ j\YO鄗" `w{/e"y$9Y?Vˑ$˔-;ܜnfW},VEZu0]0wӊ`)Hz&>l(asEc/<)Q ;F~6pEJlpQ'EʍH2 6q8~,JՅߛTxB`='.8BVdm4lqݙγ6ՌOn1&胷X͔2Cg@ydw!C52TQzV_&.W&m:n, 0~+qL/ُnArcՠҜXі||N}FWnRlw-~-QbsbQ(z.sq]":|sBOq4w!YN S+]Jf[;-UQ1fYFYrAГ Fa֋R49K(\R8q;Ky,b!TBXx,ʾ[= P,,BW_5hd24//yt^ IdQhI>^`TGr^0(s0Fʺ-@P? ) lJA@؈=fl`XC,8Ϧqù<Ԯz@nx5ɤm J X&u܀v.̳Gk!DqU0+.xxrG)Ȅ51E"7Rq$ Ǭ" T't2Vg7Nc1`<D""VFDq@ĭJ6H8LJ9g^.ʨ$C.kN)VJ"51r7N $Y#A3/I'Q&Y9b8D..֕ Oi2.\=ŀœkiǁx RG坟63+d-*n|pȲ#!gՊ[shϗ5*KZ2נ^ahIFd$9DAN_@~Jl29v)Sz|18BeW+&~G|ϤRxpRH*3]/ Z/2,1LJ:sap 32q8Tgιw>i'S**k}5X>^.)paԗ7xL<"Djz{_Mw6}.+?ɇ[)PRLǀxHiÝ %є"4(hՀGA˝0Cr ֱ`2yg!b))f!{9XpBRbLEKǒ#ۈ("B+'myZۻ)/m/Pӏ=3ج#/xQ| g 6GJ0& a<(̔K,G^X5O͗=3 KC#^pT"H'y瀚Ǩ9:r&#!AJ`լ`=zZq(dhP"E=- 5,!+ ^ kmtYEZ*0"h!1siUv,r%OS%Tvb[{ER)8B.ݝv,W]mlo%<)- e|d r3hD݄zwl6{ҵ\;St ,Rnyoc[Ĭ?0q_|њpt_vqoݩK7HT} n6vDmAi )#R5=%PR(iF%0b٢2ݯ'_޾fŽ vV@2Vꤥ+ dYihtJȓ"X&<5^+yH˫S9 VM7=+N$`]^xXu;HU[vro 繞O$:9e?ATU !$jT\pG@J)*}PYgYcNb(JUdY`Hy!a`yR) g(j5qH:Z՗8|9ŝkIXp G"QE*4f K]JHA@zcxgRc"QC-Yre[&Ξw d)[J %BgԐ2p/'xFDP&vAL,mp٤/PЬ!R[k5d!1MvsL^xSP1.uj\"EGs"GR4Ӓ`jY|a6:%K!.dv4/H19:GI9@̅3ΌelllBYS7Nv ]c[epbаDˁ+G+ 7/4xцFeȍu4L5>x J(==rƲA '#j9k! ,i*%#jAęΣ0 'Y&hX=OrƲ[Z>rvgMҕy-y!WYZ6N;z*vkvVռT( cyc_jAYgiUӥIf/ {t_Zh^ą_ajq\i&%~SbddI|ȧ@`Y;'wKޮ٫х6/"](_Mi>s{wYh&o4] Z煤{,kDSA }}_VOkFvw~m `wG$|X[qev5r>H<uOg9kȎ.fyR1ʓ@<^ N J&q 0xm*ޅd!N]Tҽ3*!\Ii>ұs(]>֤ UNֆirIz!&9y2k%5(s(|\]a8?ާ܈Il'FY!q}ԮPO1GAn>{{%կ?٧iOtm?wrOh{ƺ彽{s#d,ׅY77W.p+Uvö1yhթޓv.ߺ&_?jp<-"C{'D`\†fĐbpo(pz6P@Yw1()3r yh*cLhQJJ:@ j(i_@¬J9/V9wWe!6jKfyS (qglJcZ`rl \)R[U7V9XB_ d9:gUgOgqD!TґQj&n g;fu}wG}U~b,ȳTGMBJzi2Ɠږ8/nyzUq(C#8xx#Y .rl4&Ù1:h.0(IkK99uӟHkrAZWRȫruT,pBcL{?52k?&nI@?µ%nє}ǦOf@?IW{ߒb֤|O_3;^J) 8,/  iyZ65ҏHmf7n))\A<_TlMEԳP\!Jh({E UKBF6&li1-!LjB߬i35\jtsㅇ,m` tjXNt!gF E)7!69. m&Yo#+Yj2i"a\JZea۔.},9-H\deetD4WM=@Vٸ*uF\ 5T‹[I _ٙnu7 r!Y喐[E[U8+5ݙ/5AB.o{ ueҺYXgR3MEc篍kԶZr>Hww~M~9[}tG] Я#x-yUs;OoaNcС|sa yKW rbnk~{ 't:[h^@*!XN>q']OI'᤻lJgښ㸍_KNĥ4TGqT/Sy8Qpi%%%Q> /.pwiB.K˝g emLE/R T1hϮ:2Ѓ:DLhFI2fiO^J|hp.([]vv@v~|:~{j8"4Lg~k0ujVԛ*wgb^^G*qtu˟F«dԜ~uy7+]U9yǞ3h}A 瞇hăRUl]]O=XYYSrunϮ~ƹO7WgF/y'FIEhh} hL"MFiϚdQ[kQYWx 6yoK^ˍ+ ^|ӑJ6PI%x6Y%dRmY :|UdJx2OǕAM|ዶl%%CN(%.聽et$6UkcvBӴ䎲rm+냦?O {oCD_ZQ%Sf=(]+ԦV=WĨD(cBSv쵲# פ+6)7EV!7N+)mT^!J)KQ$? LXXFnweWז[ cjԹ:9 f:wg-G5}~&KC*;*Nu7_~z7:o#fѸz*ķ$@}Z4xqxQ ^Gϰ'VE w*~uVk^L8]MёfgUI>ʐHxK!-K~nÌOCA4#qHw?~+Ο SEYKEO_O'x#C ͇y׳qq[^27BŸ3{Kn@Z|qdJ@Vuj՜Q5ybO9'4D/oK崿{8&3Yؑ>+ se1>+hq0M`')٤¯w$<2JLB7-bc"KDItan^}ba칝]sSGXzCe0RM:J&L$ l:lLwic ͉0ƿnGk.lOcP7_&jWɾ2 l=n;>}M2v(,vࠋZJ"or+!h;ʚBDirҪ! ݕr0qIlNv+emyiISWM/O>yao,fMOǯfstdc'% b=dN`2-awS ɫ3ؔw坡?n)ķ1 {Q7oS?J`mM8qQmjQt yT돓RݢZZj}Wojly]&]8b3NgAL`k;;"v1DձUIDha!690+l .u"#EI*@1lP*J%IaIyATa^/5m+KݚyKoiwhׄ3g;K_Kܜv%Eyj[Ԝ!;{Gch!g@Z,6;{Gchw4;{Gcؖo.>$'w0ciҋdr*2qF.e1J&Rvb&&+V Y !ɤ)ZCd1-f\ GgV@Nsixׂ ^n^򴾢ڸ >ĔN(lbPg@diT Q54Fz7FШUCjhԨUCjhT Q54FШU}T }ИZ154ИSCcjxi;mPcjh-rSCcjhL 154ИSCcj#7`1yjZ>o[u@ 9{M0.?R~;H,i[v`g_;]yJk4X;gI̞ږ,S=a3RIEtB֌\d3EiUb du( E6y|/Sd"Dad5*KYɁĹMmJuόSZ#ʼB#O27e 3TdvaB]Ҕ^xyAPI$1i I'$,NeMPhC2T+!`%ē @6aLKH/ՌA,YsѕWCmXJ3@&\[vk_(<٪0A`9Ld)8mKE!F.RHFM4]{,m-'݋^(%ϖh|x&La`tgGeώHuE3%D yTT.CSRog&χqͦÛj&^F< ՘+znF*˞" A22D2ʆ'!,:şR>[YTBeR!b\ fFg]Ao]m,Hؑ}>ϫg6 {9W?}7|/1e}骋7Q?px~6.|`~Z *.ʟgk u M_=ז[+4zPC錗2zZ ࣳ]$;]RtWaId(.>y]_U 6@# D#JO#Hs 1֒+R,C պx"uYol663 MЖ/' ) y(xTB)W[VhZ %9ϫwoaHn=epvkDFOJ2L6F~_EFtE.X宴0l=d '.{_wIh׷yɿc\ELS{IoA71ÔjMwe \Z,yJ{Ԩx]%XgXt%@x\Tlz("**X GJBa#5eLdD)[͌um! l g 꽝_ԠWӃ}'t~4Y't6l"DEMkbVVS6'dR/%JϺ$Du~ Y ҪMm]-0>@_xoK"á %;5rW`c]mڶY`WxU,ƲJA ѠDa$ U1Aw2Aa1օ122(irlk2di8Rc]dة&p0qS?9Y XǺ l]"xI Tc) c9":첦l*xƋA-G)-kUa1!%+&L5OIHUn28h\|@#I7S]u6Kֵ8]f]\Y1֒B3x=Q"  }^簋;95;ִD=> [XrrNyo4Elo#Z]j2M"!ɧqwkпz>@ts ¥.z:(HeA,ȵ,ȹ\U罾M?jkKA[wwж$ J#%]" @f(PQgk렠TĘ)^E*oUE5I&!  g1e(RG[a\M4e1>#}^duէHcstHAC`T(OJmBNzF|BY:gӸf`(aA$ : !J J$LɕhTn 8%H&߼^Nq̱Y-Β;>Z?MɧUZo;՗QK.v6rb B#B%}1J^)`cZȑ"6_E $;`~ >c]l#əq_z![[ME6zr K:ӎ&'=+EO'y5Ƀ҉ iD@BQB'K}lPn9HD+ݎ)p݅Ӳ?^Kw"HZO,9aRL^@(o9c8HfcMyq&!*X6CCUmͮ' "[KmM]Ѝj\jMu{Q.0EV Jϥe.[D(o oHRZ_{G\$r%GuG-c.Bww9bпp;QO\Ǽ9 90Z;5jtG ݎ [&qGyJXyg&ǼMo<&8(]1 p4]|i .BD×ffGiJ0pQg'j@,Jɥȭba T %!TT&B :[R&٩{͖{-xgڐR'-Ke.0f%q1WG, :IM,^^+B?V48ߕhƦѱbG2 w^Vl )J%f|OD\Zq[ȴ}S:.F>? ?R^Ex7V0.Q)\ EMf(+&]Y}d_!guϋiӋM]C}B IA\- m4۴.|Y^Bޢ Plhb^GOFoQs:pv6Aqh0 #zSKȕ2^xon;Nڨl&RRAk !EfK6 ,dGTlB4hVBlSSUZ2RZͬv*Ș8y#&EnkǞ5|  ;KJG5xe,BI*(q[S1-Y6vV2DʜB뭡#:&4GMrR!x4A>Oc ( rR%}8A%;s8qPZî]. U$PNJq4sTt*R uv/3ۃ}OE S&031IC1&mDPJr3%F^o , 6N'i{p3{m+tլb.&'Mvuu7$N֤@GtcR}!6IN` T2@%vxKãwV'˹qr,f1B@SڮV9s8p8+l{{_'c:4j,H$#lxr#\GXR#z)cL*+qU/Pkv^\*Qv9++$\6KzWuf+WWQt L<"zeo4LFZMϩ T\T #KPH*".ŤZO۽7v}i&}.Bz{ 騲^Ye踌3 r6/{Osy}2k"p7>m:QWZ~BKl0BZG(V (j/(P;2 Vu(R0Ӳ=WDc)J/P zURN\=Cqe '% 7芨5\Љg(`ZEoh\%ݬ ~Q"O^-4:>\$fuLzh9|%?&Pndzˏobk_EE_g/zK^*^<4O01}?d NgAq8 T@jZme9:iNpz1&)>my&y`tƀA7AҩVAhNG$G\Fq_ˏ[c^YS3Z۟z$h[PJhКbJt7DBm GӤS/8#0jj3zWԴMp6Qcz]}㻋Is20 ǓJX%Ho0t1ө7_4x[5#)}UÈXr6Oe)>^z֟19Q{$Wj -[u\ӌG@oQy1Naqxo8(7;UN5ߦ۽??Go_y|?wosǯwщZI %_JIǃn 럾bh4lhjun|q)(6.m^.ΏܚMJY:o.NM҈pq=8Ml~6M5UNvYEx.GVx^*W~ŝU4U:߃t{#TI(i@JM&H5h\I|ީtJ9OHml/hpVLtB&Z+p:HN De!]!ɼ{;*eڞبiռ[nfgǓ_(Lrm%'-K$@ab zb 7X+e*EF-cn3֫ї} ^) J.Sy)\9~Kb Id5H:4eHj n !["CzNR+TJbnmlk,b.\WW~_4>{쌔c <^פ-WI1-@)b[t5!8*XjI_|Ǽ0ol~#ܚccwbyv昧1Xzo瓲{ƕ;uIJ&ptJ J9+H2zu`^uu>6jc\;f־V#36siUA\  5Vj>uL w ëOW5IxPW\o)l}UVxBRL*i.R+tZtGx*ZG $# AKi.8Ɣ$TXU &$ϢZ8tPxVd\q!¼d0V!rcJ x  (x8"~9 L9&XgNŐbN)1y烈y¤iHSVW єPt% AvS{a%*2R且t <AEL e8gz{p-zÆcvDyQ₃.1"q.\"EW2Z@|L*7]MO+0NfGRD29,0Q\8 A{#Sc;c4M!'gl[3l5ԙY5Ͻw0%M4 V%{56RHRPU@An!JPu)xf?9] íGLYGx-7ݑ1&utV=ODS>e61x>%୍a1PG WM3y%39~].VKM?ΛwJUE\k:u=!ERrH"]4OOS7vs6 +d݆A"v nh0vi|CMp56 rmw7on4~u@?Gƴ}E j)7| 6v}` a-j5#oe]KغHu.H%r'M&n]^ 7/qzsSSOft^.pbǺ|} DÙ1pfrX֭괾ª_ϟ!D!e!x0kSJG$"﵌FMFS>v3#?uGu2Q>/r%_!î j,dxIP(+垨ɢvTC 8K`"jxTq8>1I(J1:jDe֝K4\OEpô#S_N] j6=9qqu24~:n"nL[yKoul퐔l$OެVG8 I-lZ: :{s)_+6t78P(" lG RQ[M*Ffd j0D0R-c݊Rb iơe“eҷͿ;a0;TfmY7 *3/P EC"V'kw K `tciOSҼ* C'ZK(If Ag=t$LEt!"'Un\mxWKٴP-2[mQvA4D3.`IkA* EL0EAcJkZ2e iU'!| l CkG:Hc2lmk~9 XlqE-,XmQlTѥd>)s`k6CTA14eZa}8pFq"KFb3/:łKsH\ZIYruqM`gU.3jUf],vqoEA fmP)xk/"%yFypP)xι@{ '0a BoyȝXd{f\"Qo+ۢɞ{C2Wj :V;u9|= "SIR*0 <*}FAԿˍ\d/ ?-BHg|xW)HÑ`.pCJ+f=/!XH/eD2Y+냉5eDDL ` Xy$RD3ٺ3sj&otI0sexC;`1&Z3ltچ7oMȑX-NG4$5`8cy\7nz&\ Z1l9Б1\@\J,qq¬knlN @QQJ{[\V}Uܺź-:l_8$0S{ܞݭAhNfK9xZm"A5KoiR *TLTZL52Q+ '+)KbIRߵ߸#=i4dʴ{^'1] ƌ!Hď5uNCSZSy`wiDEsF?F5i_2mfut4 {''ċC>%  [Ω:uÃ[3%~Vk61nS.1$m)mާیa*KXNF׈wɕ3j~L.`HɤĬ]"i$k J1$t|L/qs-)jwjinIob}mqӾ2ݠ> Ö9qZl5~uKZ?mnl4!st-{mKm"@cV{GFRo+1mFMTvΡE ?}a-bo{Np˱ơ-f\HRaR5*ɝĆC?a;TZNO.,FpLwJſsdE$4AP+0;g84U k8i+!q/F/Ku5 /KIBΦ2 !͋ /V]lR+~VZ҈ϔF,du5%o ̑Ry40+mCA bR)TvP!/|TM̮/˟L**Ӡ,zAU 5V qX-$/\;=7 %2GY- 6/UCDܙr4k_M}̆vV\jͱ3#<̽:̩8s\D~=6tܯZ]VLGLRJdZGK!1f4j|5#z0Y NVFwǮSpc*;^amO7a𸽀1b`8Nnh0v ?Y> }]Sͧh08NO}gcT([.?x0?eؽ'kIGͿ,fz^=L(TFUvx:=v♞3_ ݸXmϽ?uT;['GsoV2& v)]]0^!#_2VIb)q8ZB:+ 2F;L#qŨES`ۯ(t䶄EVww,Ukϭgͥ@2oG`9ӄa*-濌rʂZlwQŰLBFQmϋ^ 1`H)&(/!YA1!O Ȣ`;krX+"D4f}֝9XjB( Y_^~U//A[@3ALM8կ\:,!He\ VK'HԼ |XKY.p7\nmu.嶺Vr[]nmu>Q"KZn^⟗r[]nmu.3$1Th! -D(BZB Q tTd k-~|ԸS,.OptuoX֔uF eZw׌~!Xij}DI9uk)ezfI:'I/ @O? \1R6?@ H6bFY? V_d˯ LvEd]I4 PDrYˢLB'DSÇu+,H%BJd di qQ66/.#'f+@6ːdie4mx>~*evM6Co7Eѭf\K.hNOv#;'SϷϰ%e`̢&,d+/zy8daMyΝUf&L+6 L sl(VXR!Jv\H!/:*/.N#FGlŁ Gc=WH&:8):)YsePTDp,Vb^@3'=PdO`oNЙ?*i~6[Tpjo%LqF\!Ч6UͦߚAgL|D"jVE RFJ,zb#}e 9(.(sP\D栉) FXT쌴9o0S%0S u4D{ |V[UB7zf!% 35 ^И fn(K!dVg2gv f uF$ny-)ɇ9u󘇸1 *N̎W --JX[5SRaPr]mo9+~-Kdvع[Lf!,.dO;.: M:)HÉ C4)"xD8D)m,2-:K1$A/ŽWZE 6L IT2Ve45":d=*.hc-Pҍf;>; ǥu'GQ[%A 1⺃8I2A F\?OnIS-M(Z%< #I 3Qt5r9[*/WAE?:0U'93Yvtr@0TN%9ÉLܜL'q`Q$:Rzꀐ]$Vyh׳k|nyÇ::!KPΒ:_GWǞN~7KWH}Gھ7K~~}n>(L1̈́8&wsMp&G]: ܌Ϥ }޻Lݽ9CT̮X_xs5|plaJy̅Ar{;Kp4jZM9C[L dHBm>a*mQO١ɼDOn_N:*#G?dۨm*%:Q8בQA6U(͈=_~Bq3чg6͇޼}2gW? Lg[YPfG1 #7~~м8|hn.COL.㚒2=Ÿ3ۇŒ[+罟7g\V-m^q^ j~1K\OzАa+쐮aybvxIh~[i:qF5 6[P>('oBptlRS RIy'bT1:<*j#=uayU=c qCRVƠHJ^qf %d19^E3D="{:8|{|hL܉3 \W{͙u8ﹰ=fg3vʭ;+3v֘U]XaKjsͬ$崹Fv)1m[&豷F.9!]o5F@5t"GebG+C2Uhe<Ձ~wzyldIDaלjw$9E3hT,hgo)!3@G( Z$ c"\Kf ZstLI\"jO8KNPcgW8WAuKwuqPcsv%n{7_<1 ,uLd{2*KEd4&!)kVi4'z)`YVe6נh5OSO׋-[:W6BU\jDzˋ/cQ'(m"VLа6y 8S1 Dr-5QU(X>Ej@e24c@! Ih%rfF@aE-vEu9T4K.sF bHcFPWC:Q;Wk~ەX|싈0";D$4:oSGI(Nh}C6Ѡ I)A@/[JG)8]%(ZqE0"1T@VJC𧈯#.NF)YLJESM.n$*""X[H"RN;}cţKIǞxQz@ qK4^kNԏ~e=fiuIp]se?"E#hphX=; t4>mxq|~z9kĊ+r)"W1GC^K0/0׏Ao0Ay\s?8Ȧ(L(R-ko})A/rptAR)Hb)Cf"N`(bZ D`eA,ˍw3CDVZG:i4·\1N0(=)?h-P/6>7/eݲSC|% ZiHn=;dU-ժi iXaFR56Mœ2fFZ%N\y>IYOB=!׆<;<,ʈDOH-㑊)@#ZDC"f溃A+ؑ@p>3r|MBH H&:X|J AQdҙ0h{{i}1үSֻEj01E= ޘLJxV$B58UampRJ6{-q#Tɜ/5u=aA) [t_ھCOm~Z?osNyxz~MO1 M殞Zi;NU{~u;mqdP4H-9{F>ܬMr Ƒ;Q%v"%w!oNg/ av²^BR5F&Qkr`.aJ,ߛmJ?m ۂ= KYwoodr0opRLGuy\n#doY6r; ah7=٠%Ͷ8Hho:2aal @B-nӖNoNм{t.ucwFRst~6eP bxOMyyNgsn6 4sk1%/83Q5;0/n^xn3u{KOeT*18r$4u"BLh02)4>*Ym82 OL,F}}4 su-$,WQq;!S4xgQL[X¼ف^c5ӫcPL6GÎvYKXqrфJ T#A[ic+YUC -p`SJ+]pXQY7wּ昏"긎.4A#=yi?^* ƵҒ2J9H2D=$3$#.m-jI  !i녥(I!FcJ`¦j1vvh0ʥ%zϜ4Kl/o/>D^ۊһIJVz,+)Mz +\A4 ZI 7!%M]Ʒ&7rg<2MC)XºRό7AIS7) re\(@wm_n!@x8xɾ N_b}JZQLREW=ICRH"kf]S]:9%y(a.!QpjgV[ q(\kڄP UQͱ  p* X0"\B)A%1Q X|-*7^Wcہv]ZcyPi K!N[И"z ' 0R\XICvhԆ¨C~/֞0agS%ׄp139PS%E;}{Jޝބ Y @ڕ7 +d.8ҥ FwypP<>(F#)h0)o/G!,_S(lG.8f;jOnB|;*sT((eE .HHș-Re;xUUpHe6.\=BGugK/M$OUN.R%i<(BOYQ\T=;_T}7+"(7GAv򯟲RT%_@w ӰaܘoFSR2P!9kW@s"uCXL-`w/\Yd土Mb v001smC:]˪bk䬑@ڡsDފ_̒86P3x$煞98[վB;bk m6wrs씹Q'CGMPbmšc\gRqZc0dp0@%-:ҽʠ: JjW_RiYztIFڤ2U{`KlQe!=u6 VQRƌƁZc#WL#,8ZI}#DkKMᚋy|(rz {~ rRކ?L>d3?ÞCfyeI.d|%՗`yh)ǥ.~6AcӰNFbPPyOz8 /d4i 3:=@*+ ^5moz/4|f#zє]nQ٨lye's-ZxXGΪ!FI %J`Bgs77GHqG@s^%b_+LyER;4 %\q W#m@{{qO<^O'M]w׵\lG۫K9=ssLI2`* ~{M[܆b3@5ۃ7NĦul.kZKrsxޘֶCvCm0@wNߔ"#M?-kI,s'};i}>帏h|;6څm2qI Mn_R}Y(:m-,1y[E2uv4@ZRbFbk`\[.u *h=+I# I+Z !D!e!x0>՘IDk1h`[#g ~l &ETL"e5qj._%˲EMi6~MWM'Eb8+v!r)i㡻 E$ 8bx˫Žyt׫68fMvgn2ܓ!%kn|kpK>;),?*,k͍zI*)kï]7}suk۷~r(>GR:)4N7!s)Gƀ޸@%r%(#.PJ;JnݔAZL3hÜ,)!(gv,Μؙ;s⎚1`󩢰 #P Sp2gƄhr[5'RΡ[ʍ"O]֑aYqG mR.k%|)$lNlDFJ$o=Ӗר:ma`>4҈ƲDKy$ hq$7 3k2]3|:(6ǹPlCa;Pl#Zޙ،Ԑ!eܻ~.(v2 m0#jɣF"' B("€$[EUoFΚSi!~:OgײТMQe]\QBt`p+s[P=$(C])xLGuٶ}.T<")9IEd?\afXT46 KXA+@2Z?  - C /Xb`usZ08v2/`߁FRw<4^:O`?ִBTB2 'Zw^D ye Q=&n!SC.)XXZicc$a~H D, ytS2̤ͤ'^_~8~ kjgI2d:LM?5~ǔP߷'^(ٹxx!8CB +{™*2b2" |9N6q:dX" 2r^3)(8\x0u^ݫ+aUiC(XK qr ¼]*i0tI~.mvm !R^ruLh \Ir&C8")Z~'y-#-GNI$@K*`䑕1JAH}WG!ϰ Bc{W X X$0.ɗFZ%R!) g8$E5%ʞ]{zY#ڝ$rA'sʛ\F C}$Jva ;gz_/D O @MBGu&5S)%nvQXr1OCQ[Fmգvon % PMA R- ")HNfx;4{hRQ#քHbHc}PWCxt0g7ND2 bq("ˆ{Dqۊy9TC:O+罵9K4Aj.OVJ!BxR PB}D8Jifsj8Z:XgUr(.Ba\{\HTD8!4Pǃ yg%+EV[>rCipcxΥVǁxQGD,^{hx ~Xl{!\FQ?g?PfRُ㯣j>f޶~F_yV j'GD($A}㏙x|peO(7Ϣv'}MsP>w[\0#\vF$pAgXq!ZF(sl4 xP*T0-d1q .D\7\0(3Y)<k1b9qWדzS:%[~=׏ˡ;iZ0n;UZrdb^ëC bᛇ-`[ף :6 7L~אL7ḃL_1%,?Il؂=.a%o6fk\xxrW|KS2__u\{k PIcځrfbs|esEMSzTRv͡喌)ȳ=>vRY` fqPyeӛdL7̷+gV_'õd'kgxu=w 3do.g?-VEͽk6Gu$vMc&oG?l7fq!j3|?<-[L˒R?. 0?{>x,i\N_ni|7 ԉә7lnNnluo͉K,6?\$?̚=v|{qOR_5s+H/lN[%L`8-}>Z5{|!u'}'9|W|{YG|%S~ԓKs%?+Ԙ3IEm~ݻZ8<9"jBu01}-3EO"*-ĥw;\xt'ïA9d^xdJyYW^;k~ێ,hJstm<ްkl|)V(ۙ{liͶzQTFHݲ5\ɴZ}&4FSJFB˼ Xҧ> e> ՍODNQrZ 1 Ouz\$0'7@/@>WuT4C؉ O^1Ś/d/DvS{*xr~p6DT!HDV.R9e=Sڦ.mDcqȍɿ<Z:H1&J)(qQ)1:,D%j +K{n/,Dkt2b<&ZHv3yԦ5ݓD]?Q|E74q<~xI+,<4)aN^Ԓ\[ߥ.E]zVo+ܻ?vޜ7.ۛ_x;/qE[c#ۤ\ML"a u>hI,oTmsz:7AX~gv3M= ~o&uu !*w\dQ<E\ Z? ^f^\I4FRӇ+ՏW?=zt*! &p_WU\vp;D.n+^ fP"ˠ~]&Bϯ]ϵTxp}p`(j,G$F [)_5;mb.uxFtܝqgpu}cPݸ+[X ,/oxj{4f`R?M[ذй,OhB{0-zx U=6wi#|;厇YEw#s_gex9"U40G3Bq\QdR`r2Oޛ66-]|>d֤Lpaso{;m.%KsyM*!9r M7H.(zODЀ>TlqtN"'8Mq GW2sㅁIPzs|v&Js%'J;7J+ruBVhcʀjb P!"e3щMntJ6v8 ُ2]ZJIJJ'I(%ZdD-)02Ra#!I\Z\֝XKDJ2cL)m$287ykmȲEnڪC]$3@f_vbEEfg+ZAC# |hFIC:{OSio>"U-euUQ1'B&4x_ ݠysYU-wTM%9ѐJJ-At@r=ʜ%x_u}d St6G;%Xr!xQ+QЗֶڔ^*QT3 I:'͋Eէ (хk§tp}E ܔ43HTG Ht=+ )D!;$JkCvYviF* h-͡]K68U& ' %jɭZ<ĪuUJ|Yɠ-],@8քBGWXLĕd Hqk fCFk.4˺ F@B.Qz$ic!sPѪ J(ڑkj $*Z@ԓ*e ,>lG?5aXU  JǁԬ˃̨l`WP&4ɎU%N v ?J V+dWVb@,AnVt,]\?7(c:@:om0S @1ѽof2n;clQ,8UFݚ ]ZK{^zF?Lӿѳ[}2ZCCIΆs+B̶jhԜ;%@NEW7ЕtE(_wtu7j3ٛxvB}Q^CC?}y;H9 ,8TPBX n5"X5#&1Ά7J9&zE%4,*wVL-հ-I͋V1xezqR_.YW8}oq@[_<и^FۗvpwKe fcz16o{KhPs@ }U\K-ԗ?N M;绽Sjϯ"F}xo{34}0E! Qٷp"/3ʶ섘M%2%Z#=J'8`9%ԋj.tEhv"]`ǧNGWUUQWm"uvU6/Ұ+:G'Gw%%~OG{1ԭˍ'M[~_Wκ~w}{cHOx_umkwU8 |< 2XTd7г@֯3Co^'z޼p/^>X\ 2+tСﶘcfCW\ ]nJ)-0] ]):Sψylj="u1PnLLW߄443+&Ά`BWtSKov7Qi*gէtEZ.wFmF )y(/lG29on<(=Y,Ϗ`M]+%Ƶ tNe]jT'q*ؖDZZoŧ'}bZ膘j2kY##;.A[Yrjj0Z!zkJEW\佶sh7)Lp鮞GIaBͶKaJA)삈:̈]egCWnu{\ڭ_h$w|1.k3Mrw>Nֈ_h$7"]S1H|@F]ZØ ](Ӟ^磮٬]Z{z z+q OW3ս>H~h|?~v{Еez3+~Z1"^l;]ʠvR+mfCW׆[OW@fA2B !fDWXϧ$f6 nP:.w+҂ߑdA^}Z!G-wo?#`glh4z4M(udA6"{G/ʪafaEw_'ⷸjl d[\-[IhLζm=3+N͆nPs+UbBxy'X?SW" /ޥ1~:=Z˿q 3 jom>~azDp~r)ͅ7{ߧ-8aWūVr{lao56c`/|Ҟ(J;]Z՜<}>󳣂G4q.%moz"a&2co+/ 6dS| ־=Wj_i:/﮿t _W\_m(↘Bv-!;@k}O7?H<=IM[+if컼|`:>gcoY}oַ+]87A|¡CxmE2Od}׽yW+mS?'aqxp*+SJD^uD )6)bN٫RhX8>`?׋zp3ϐd>,ѼF^廤;l2Ck"6!:7 ]yo9*3;0Xdݙw`gTR$ىマU:JJIX#{ vVk/;^sXeW䤹OR <\|9V=l x朙jɲ<̻|0t+3QR}7O* ҃zQJieuݎ'KuEl =6"Ť&3XcnMU+r87M(p^fcyfo/yt~AkMܝZ_vf[3~Ү>{=@vw^uqj =0Z宛jWq*NGX0]%`nd oa4IUk ^N/;_8 5h[> `=hGa|lHe% 3Б\}FI[$jKa#B4%X|M;yy󿵌ҁfVeNۥݧg?D3AUp;o8 ?3@c#WL#,8ZI}-DKEx—/Ka*oNoua_bB3|ʥ=3 >jv:_ k|^mrn~o:,h)>50njf\~?*h|2\LI+9&$C[_ʟdU!3V?ŁyG lxa"v)ĥVCq Yi_<& yo&0 IΡzCHư1Af+Hwx?6,BwG7=lWNE:%Eh9qkKo>J̣ ;e97QHӔb:)$6j$$R/ Cxd!RԔ1тFQ4`pHnH9ۍhlF.t/!qnFnn׍ЪYN2=>5AS ߷Ja(*jJi繐J) bňF\1kT*kc&PfTetۨQ1,%! j@Ca ng}r<ƿ|b=bX.bh,bvmf*RAKoCpv *5:Ehn+ZD%\S@PńL*z1Z>d䈵!0dR`>]y~46oϹH^~Z<5l =|ݧ~|vFdxm)Tn+շ )lLD3l,,1XE-j]KL]6"UjIJ5ZMS-:Е h¤:$d&#QHYu2LwZ D佖豉hj ih!9;˵[p'hثJ.e`8yF}ɐ_=u.'Oa%b}oC ?ζ-UvݫE_pHgv&r#itrF&UZoԤ dƫ?i3(]?қ'[*],*fܺ{Q+wRr3?$`}}󷛽>cLf2OEȪف(pBqOh'<8YYv E69]N5zfmC:h]G7Z5achc#jzw쑵&Kp3"اZ*$% XZYh645 >{/jRh!q-wth>wJn%*wvw"OjY|;üLPp_@Rf}*>y3*3?RItJ:[t SR5e`Ftك2hDN#T 0wLe2{ҷ? a-:]Y|Xk=Ǝ(͖I,z~Ii%Ul\L5/Vh8whq$3 bO3ɴWZz8+{wX#!>7^ԛ3OXSƽŋ d8ra6LQò&'Nvi BaQ9Q0,k2 (,Y | \Ku z^cKrl+ƑN>+`CC-%[꺖u͐fm,@h}ȇ#F1ü?ɲޠ79Q]U6:d]u}Te_'2>]v:Iΰtq mTOVV];om͇޼}?> 8n- $,8~OOh647o*KӢ^|vU]i0ɢBŖ[;?~y{\W(e*BR$rWZsOlA7YU}Hz{n%D%ҋu EuoSFM2*x_u/XGr0]`TK(E:HGbSZjFa A` tlc| z#́6ӧuHx!4^z)GLprCR+瞅h"HRJFtLvL}1X3 rI'm{nlAƅC6ASXƵ$Uw_ʤ\}ݍ~j"eрmPpYJL&0"4q{EгxݳF0,0-X* F A v(D!L!f:\Љ`M+eoB( e#MƒcIxW{ 8D\Z'uĖ+iPykN1sD9_ Vxzdw.f) P5EwtU/﷦_A+ V9~C%`vhg> 贬/}DdD)1lf%@FKpȲ6IZD(ZjCGj6yFu6X b+L_$zF`"5xG4W(ҬVzS=\ !7Yۥ&rlo =~qHkA+Wٮ[s*胙/[ sbnMJRnjEqvη&*%no}jIJ6.+ce*wgOL4tb"P5ˢZF!Ke̳艍`p(8$gBs;TwqIMPNi0E?V#cLJ`()nBN%gQUGCǠ*nz̝vX-@&{gqoR0#4'(xAcÚkcl\g9n)+.T FzsOw}[.YT3sKJxy෤ \0˔sӪ'R& %Tk#G4cf&0^zeÒ{i4.19'Iy=l(1Agk=]x<,xa@샇kC~ dTaK,1"n8ué/&'G,6r:>ԧRA8o{y]IRQIs)SDhZx@%q6AE(ƹթU"LZsD S}V!ǂAV M.4J#c1v!)fbq(Ba, KeQX;wGf~>u~sCeGhyoؖsbNj-Q<NtIʤ<(oO^k"63r|{$>Ԏyg FXW "rauR 4(Z_LO )}Ek01EwvCA1 ,d/MOIjr?8$L$Ł%&Adv>yD * n&sA_z 7iDRJJk(7&$O埜Vck6;9LOSFI\q5)_7aы?OB25|'eSLw'EN}F$nsu f9Ԑ_?q+).’궴hC3Ϸ=6ɳCɫ}k؋|3?#>0l: %~fH2߀ 21]ǐ~0۪$b KYV3OuX8$mo_εo&$i ]!6Ó5ŷn@4ȡGL*e^nr9p49鴳 ͤ4\RMh]6ĵ̒5IYXoӿLðy_:TM??V5ߋHV&:wG1G7ÈG?wόbmLx5{[ }쟞>eMiiXEaqrkX q==g@{FŁ 7YRw9r= a}==Ղ]2;a?nrpYN4y/#^˟$2Ҿ͉>sd,:h¡ݓuE|88o>WR:2L%wb;~s:\_jVQ͋CUyhNqXjFCP!LޠޖN,,-ĵwSƳ479ޛ.v:5w7R"cڻv%1'짼-ŌFLtS${GGa˃5NYO63Q9;u#D4G@lJB E>v4\FɝZLAG'Hȼ!ArXtUS)TL6-\NIGYꌫ*=(JOŔ9))!YGq2qpyvQq6?N_Ax ׾U>r6 Жvam3к0KeK5*l࡚OoGy/&e_a/Bs~Fo>C`eXӞAU fJ?޹**$TF ̊hXIk~2 Y$(H9wX#+cCrĖM3 ,RR]:/tL?=ځWIR}5Ov\ h]Qm(@8e}ێgveA_68A{7xī߬|$u $AI$'97i)1:Ŵ:J/P[X^:k֔z.hCL$Qß3&H|2 Y]bQ L\R LHqm=zMTt_r<0~oR`('b8w.}XuF 1h.(.Z7 TWዼ*`%M%TN803XIh (U`જ%_Ѭ+gW F 7lH4rg:E3 POD{iBC\=lx^ڈ^^pqq0U 㸎rkf.bs1l~> o~ܑXyR|*q/%fEFck1i|mnS{Mz.;ْ㳾66jzDe*RQaönzH{7nFMmKa>6LU^Ɠ,%D.XTڨ*\1xkMo lO}D8 *x#BHwhk'%` 4 Y7}^-]qkWN'm|mOٞ>=,GsBu2y7>SJ2-b@~S6i?=(^pp(EGbʚhzrv#jb[Zdiq.)-ÿO_U#Wֻ\T2ySqrAES>uNzs>ѻDШggjcujl2h h])tD#L9Yçb1J&ijb (w!6 RHF%S69kE3O*da_1v7v22v4%^gY(ؽq};J42Fm#4.;H';tg/GLnei˛nm6vW,gS9i[kEǍ|w]~-@:GIcNv>Dϡ;Cvw[\ݟ=d~2N~C=[;^չܑC:o,v>ܖH~اw%m6eaIlC?O􋘾L\$[% nZLRT-*/7mG߾$/K2C-˭d^Q*WS{k=JHV۩=by ]ynNWњ.@xe]Bu'ޝػԝ(e1JbRޢJ+Y !`LVDGLƴmWaC,1#((mXTJԦB;jm0']]r_^ lehsǤRf#sL}w%yiosN &vY`IV]Qu .d|,Bf ^Co)nm y8^dYozgv9ƀQ%h4bl!JPRgJ[4J3;cZlKtc@؄xxNF"ɰ58~p)ֆl&e/#K"C㍁f<ҮZ 3UIX;OõÏE\=ĆzKO;>y-wxY^1;d`K&krdA6j$4|A)㥘5)so^/|`ۋYR]WnZ.KP|;)kE6U؇BtBDcŐ,谔5[Y5yg@@#4d"Ek/&LdbP?2k~EbV$Ѩz1BZY$'D*uvd<5銍1d^*ҺޡqZIi :E U@210!pѠ JdQ,R;1 6sOӫk;uQkv.|MV4~hxluy#˩O ĺf+Gzuè̺d6lOSxtxГ1 7g];G]QݫA:i#u`gqWA#NfEAl9ņϏy:ӏ_~?G/?+_ \Z1kE_g֭Ckڶ̸-;!rܩ1;;U |qpuys4-7&[J*\ /b.US~_*<~%ӍiEL;H)b}tM{>*pͩY#3&^\m ЕW :PJ}^JlvUDXhMb5Fѥ 0G&1ho%m4s!hg@7DdVMMe`@)XJC6.`3qs\>Rv;Gu?nk+Yu=$%y8pJ%07%$${l(; "7JGk>oTSDJ :"u(U"S5 Gؗ,oSv"9!@k)(&WzZ\d`8h}$򟦼Ea-)P8,Y 5B 1KQ2HǙkTf^HYƼ8EhZeNhrNi)ń* ~ l"LvG@L6«>V])!i[ɼd,lk Fӝ$8 dPTGC?R ES(ɳϚH"J_"WD5׻c rUNKi[աVyauu u.QF3h[Li[.S0sB olrz; Y6nko<lkԻf\^IwUJhU\i^Kw*]J :cay?yEdF9+2Ƿ*t`#dg$E-rWPPMJWOrt6݌ecgkjp#-ĶR$dE'UBgl&IIӓ֬+p2nUm1 nV<hP ;@?i<*<Ӵlû!Oq{,Yg5u 쫊1=\}iC}h]zvg.-9{Q+ZFdlm\J{ pO3Jjćft18+v39& 8PBrK$uC.brmsVu!#GY12٪r d`R+^#S21  *MQ$(/|2R1'ARspEe,)YH4go7 !"SXO'9yΉ$xZ5(l25 q-toa \SE%`P% 5ld)ST(eaSd49'<6Vf-Ca͊rrK([iudA526͸J3,l63BhcNp)Q˫ʌהz8er=dß^"&*!ĬTlFԥU( zu[^P))*ljB„ v‡ [@[Si%Èy(lvlڶ1j{ W,cYCΠ  !J6f(V%3xXu7Z %#BM&@Hđj0\LF5a3qéQ("YP]cDt="iKu>Q:_f9c)RCd 9QŃ[͎-vG <_x[گu=7Ulr,9nF1|D\ͣϾ[1 wt|eG"Y-b"_.AYOz|A[ɷ& mɮjǾg>ȓW[kFm1> mk;/8l [C2<0A[ot ʁ)@EK2t=hPH(zMgaQf<2"E/ξ^OC*`c  I68?19bekW-+O}-~ftr~6,&h  KƄ N[h~!BiC { ! x'!\R%I]IJmE-dT0$jv[@SN=FΚ<>ZkLêZV_m|;Iv^d[#zJ6)bjšB JƛPK*Y`,5SRx..s=kG?xUNJ*2aK3lrk;l%  dwnك+&www{D. fA(pG9!4,r[R3VT ChU` =R\02ˆ9ѯPqVY%VߔόZM¹S] 0I"r:x%pXܖŐn%Fg~d~2xJWn#ވV_otatգ d%do&Lp\AfߔK[ ֍"Ԗmdi+yh&I&U-|rqQ!K\uI\IfA$4Ald*6[-<]5[S.3 5=o?vvdRԐ F噉? feȘ.GKmZ5oO/]T%5Nb7 jjBjhzVԻޯti2.QмI?4p1}yqa ~[v=,+FK>] cP_= ,ø _^|cqJtKo|'+,6X9<1~8ǟ4?Yr%i^EOsnf ?Jp^g;|{kJӸ8_2[lTOWwAf @, )$xy/d5عW9geh杭v*~\ax@#߅v.tv`O@!:@o} 'ST{-G+,M[σ(<D#gOrP4: !Q6.m;4CyC}B܌\䧙;_pY#[ǜ(h2)Qp҇S=ᑟ>ut %Ap9Ԁ9kðRNI 4!`ty^W^;W| 8} y\Rfevjn/JnUmq\\gN9,nO6юnٞ; ;{%\3K.h)cRZ& (/0 '2 #Uj(4&%y. {)ƁYc\F"qsqLڙ:-˥7=PtW+Wv:&C̚'|s9Mr;0.ݢkDz&^⢳&)HÉ'ˣe8tx/$$*}lGå;_2ٹlcOLh֒'/aZxŠd@HFS"Eգ2[,vrwrBqo(+<8Ճ% .6 1⺃YI2A F۠l?~?Ɛ0*XGV7<)5!.`FuRBR ԜQ.XcRbIPt~P#1 VzE=Љn&dRّfr6?k}s2!ޮYyFywD q?o͎.fnNLOCT87ʍ-#A YTg?t.tP=Y#j02pŽ.{8,l(E8LJHt6PNjORp6ʝJStO8YNc;x5)[O /W*(fBBGU+첖iix.qztvOUyZ-W~,/U_^ 6(R^+sn?i4|[c92Oƿ\K8;ƞlI=7uÚdysͤOiUφxtUpQ2&zmWɁ\,oFz{8}L}KNM czD-T-BΉ[~xq޽yݏoo_O՛(3^?83M *v/EJTV?]ꮹmV]u5m5]rCG)2[秃_.?3]\^r\ۋɃPJUmq_p=y(]alGK <;v3ؚGwʑHR9WV> A31)1QJQGހ+rVĨb:4>Gz͆|9 /AD@Ơ'DS;i93c 2NE5Dw:9b͉}.},E,_(7ByntlzvJ[u|\Z7!EH{&oi3-( pϞAX?> }2wNX0W rUoBu+T*(T$8D!H$q]U=p rpAHary'S+K RɍZrI-i$ PcS<5ƫ`a *iAq(ؙ:C/"K$+sry;+Z,E"u`q8E:afA-\AL>Q*45/gty[hRѾp} px\/ kd?B }L)-W KLƒb,y DqMj%QF̞g(Rk ŋ]ںn|iB?A[ sr6'"P O E=f+d`{y.WӸw cx-:ăFSa]pUX4F)2yuRh2Ad={縹k~fvk(nw`5k#;f$\o= @f0fFr+|,Iw<8t~zzz<(2bF$1FKxbi}g YtQR;+`շKϷ.d.hq?x$!Hq/pui]J^+ȥ5Ika0IhoYїƺZ4l!CSPuf7tĆydGRN<pP2eݸM"\ĔP&{l&4ƄDͧL&wwwpDAkA(pG9!4,r[R3VqЖ09n^#Ņ:O*K&ؚ:o = : U@ ū%bJ;N% ޚphM)8.+%guQhn'oGMȖ*Y+j_gI{H+ >~ .!1>%cm U CRH,zײ8È7/70 aF[r#Zu}߱m]2ق26n$o nln#K[om+֯^dra2]UiC% //+uy֑]I#VY@m:ʞk=tbyӚbg "=y5;˥֙\ycYH- ׷T]QC)J5q2fbyr66 0蕾7%mYzƆyWEUa6=f?Č.uV 46V]w/j3]m/#ѕR4;|{0`aC"r{Eks-=ˑ4zgIAd#10r*Yt{%Y: Ws : E)繸≒]HJFc:͋yZWW L|)^-껅W7ȫ«ɛW&\>./s:Ow֒v`G("A L.*T{Fn02UG5py@pu TyA>rA. m^O޶><_E/٭&KF +NâJ(OfYo0`Qe3Ż~*jKtTWLskQX\ ŵh{//nā~^)EynܵM-wx)Y_ em)bԥJm?HR{ecN\6R_{˦RAOK{bܳQK@BdLJꙦ@ME)"PŒrABGRKb͊ 6qRI.՞yt6H 6\ 3jRvPn1ܚ*(Wp8 4Ԡ28Kde5&J'f E@ɅJT^s8BG4ᣱ.g}-Qy Y!TJOu}ٯq}H{s7=bqJP9 cPyvo"W|{ozo~| Z\E2wEDͽr9em2rXCT;ŵ9B- gFͷt^+O(=TO#RQt_b?ULE}=.w CƅZ(/7tm*6u^fJ rݷ,ݗaFiK!2dTLZ,^Qd4OM jXI8%3xC-@rsI. %ՉKYH͚;=uΆ,{B),ˋ8GN1& Y\Xd=II8 QQ\(!3vf*0c[^(: <<)d%㖸e^`ej&q J{_98WCE)12!W*KdV6+l)y.CV\d{8l+J 06ꄺ161aAX$cwz퇃b ;ڪc\ 4% m͍46JABD\I1D5E(uTo$#J#4$(F* %" Tvc~8andƸ/3ؖ#9>pG\Y𜷁SGI!hCpzo:;Òm/_<ŕQt"{09Fň4]>x+ZQDJ# "+>rKipFs ;]æ{`ak֙3r5Y\?,XEF>X@,R/NچRʳI$Zhq<0C\YǠ~LGp(ojS2 @M$*\"c#agl<.= ݥܫV]oZfK7m6*ٽL1 9j(͜ P,g&lRVL4Xbcf4++ i|gu3`epd3 ՚1wHT);g=)gt<'-ݳsϊuբ]8hOYJ~F2vU⹰ UήJM]Dv y` 4^o߿UW߼BťՏvΙ;sGjI8;^^?I ~ gB,pdõV Жwt6[30bp̜w?93 Iw~=] ͆Fl2r͂o2r-"A([)0?r+KA|gɥ*_85Q#e+A,+6E}颂O7.p/\ĊțvHӪۢ d/VGkyH$x:RPNDV3P)ǍXOIR؇ `X%1 ^=Zk=?FҨ%YChbwe5Ů&YůU_q) b%$eXH}ذ0yU=_̫=AJ#bPNr'=wr5 &DNFZDw:E\++Qw/;*=7;gb|ܙb9`  $Si=+ڤ\l4IUH]qf+I S)Di)D0)NFd%L#Pp1&;C yD"Ȣ0*@C 4\UT3CMP#^DmqT@ dZ äbIh z9kڎ$B}s` Ӹޛ.z W(:cGK*cƤOh%^ ~@wS ްčG+omwB>M}Qrɞ<]^݂o{.T^Uv>-ⲀWȊi+U:V!dtIO[{55!լOVޫ qFc$8j]:,<%QI5όhV֫A znl $ 2I@e$gSG'.[Jh B:2nv>x5Õy=_ h U/=M>O}Uh6U1#" #@ }n @ %JwND7*ϜnTݨ<#Qyt6j'zBy\F<A+wYW0|L6FS<Lsp bzTx)FdaY`- YXBjbΝ,,KIuOq #!:JfrZxP0'ɬpe%)Ĉ[ Rnl /%~7yr= +vd}sirĠNq1jM`3)B)E ǔh:W}7>Gdq_ ¡d)tLz\(D McC#'е5Łi[2raP^h ,8 1ѽ7 !AB?Q*p(hBڹ**$TF ̊hXч<$>?psRogff\onrx=EGCdc|9|*)q C5  4q01Ax!GU::޷L13Pq3,i7g/J?ڠ8Ďx8̱]?mx5s}k꛹xyeFƴ}EjSn|lzEˣ*J'cϣ,wpmVY3Q{•Wѵ[iR]@v.x^]M ygFUc_pIry>y gZidP|NǛ+.׃fD֔CT]Է|Cn;=qor6/&z:&"K(; KMq7|f0kKV2]8MjBH-,n:$Ns 8c)U1H*Eȅ TiXL6ab8Ba[-<-\Jm~q[s_1h=MʎFϣ㧺>ŶuRHDM L3ԝ ^#tIIɣ +ciʢR@}qY*9u[%E]..vqo#Qp.pB@6(KW%BCVY}boOas.:%a7%Ml=rix ~lrlwVFQׇ?M(1ptaEɧqR减Qrb+eW Wp_Epx!#̠ǣ;-!2{@B2[Ou*"4!йrWGNIZ;ƴBl~!'Ul;ܥpͱ厈)|f=O;{YTeye2Q?->GZoOӯ]Jqv>W?!"d)fތݭ^meoi2 y|ks1f-^]bvߖ ˊRޮ 㰸zkXVxM;[ތ'PNg8*Efz܆0_ZcW_+{`Ϭ9s_.m$Ws}is pq<{3ԞЎjp:Mz|88(o#?zHp||_|۬#P>)Iٞx]zXKvPۜq*;y]ŷwr'RM^Oh+'T7h&f@Pc`f_Xp5bDiJclOmNp=~j#ӶEj3yIfԶڿd-;60m{̶,ݤёqil{úo^;eOWj%cۑ[xliͶިlJBr-@ 9A0%tׄ-Q> e2Jsy.̻ٙ}/PC7HeI .^CQ)x5|792zy΅H8X zY\x^l8J>AOvDdIJ/_l~Q H%e|%yVJ`gznLբ^ԒSߥ@]zVwk,qOЋ zcE_[6 m^frז(+ԶaP|׃6O1̠"2\\`1Q3g}77=MSz]ū;8#TOFA؅z0X?gvga)ҢƘM6լfeR(]?+3Gݢ/M&{ ~'s樴WƗ[g_Wf`h? ԙ֜!,<;] S`Háqej8uwSno:&p}AAtH}Cxym >Rʍ K9B tUSn WU^kwz}_^cCېzvz[YԳ+6Z/bUݎ~ggCnGUeg8߱} WaS//]yTw+uufyva&_5g* ~"y#yhruDwK|X?)o[4 ޖGτ93LݬD9&̤:SLR wgؙ<(Ro^^}VON_ Ԕ-f +@6ӤF3ݏ?lSg u-1C,|F˼UDm}3xv.ut뎵,?3[E˾6/iNS{ç嗯=r?)DgzjS#]>gv;DxPw߿{wi^QG_R$eM,ȤUtm kgЕ2uai g~^Sۙޭ_C^EM⹑%$K"o8:?h2 PFj=]"|!>~>{7N7 }ɮiu4k"PMW=T-5A'g_++ 6"SJ&*üj2ē7U)viMMoH۳\1Llx5ٽO|yy{E{o×mg [KkkCX[+i9ܻn{9쀀!{㎮RWe1D,*Lf3-&+d3hx[P"ٸi1CZbdDr;(ܫl8L^e jK+S󺶦^РC!h]lDUk@5Mva淰av}u~]oxc-"`njmO<{pgB qq&3^fqH 5#Ii%.:}>G;'lVe HGg GyXĽOA gmO=W;x+w|=z||;6ho\xm}-`:NWO7M7>t멝M~ׯ?$dEvHsK&# (G:o^-nWȾ]c1 jPdJQU,Fv'`^R^M=4>9QÚ I]_:s+g e&o ]BYzvU<(]qN5QD(ذ P0G=Gi--S5Z!kв))(k-*R4-K TRt•Jݮo|+r*v^V:=A+nxkR#H h45MWd  C5X+zA wIVg^#7Й'jЮ;itmKi%# dKbOՠS=NrUzwَq=rc.m{:w=VVqmWŮ] ю+Eɇsr{ɱ֪A*<<pB)hUڪ2qDޜoAwqj+ަ\q\4xⲍ!d|QW_5[ P@!d/E=d~P\'.);s-*abvmQrc~&úy'hr!sCХ~˟xP4_AD9|Ht>>$W KV{I/*L9|&Z+W,e4"bWLb'\MW΢6"\`UBĂ+V+Ub &+UL v X.F+豈q*K iĘA߾BBprW+V WoWs!R(*b?pj g|Q{N7J?O@q_U*x~H!osT'؏mj\SZ0Oeg.S0|2ՎzW$W,`,b^+V&jޠ4{ X.Xp%펈+ViDqVF+3re4 [WȄ JA]p:X6Z:Xw Wĕ5%e3gʟ+2پlQo_is} TcGcܓgZj}2&hI:*\`#0\\%b*t\mtpi# 6:\\'cB+V)Mq!*"\`)M4bbu.t\JĄ) Bu?ό?S89z!mFsTBiIrV݌ңx8&qY"ɴի\( .9U0![ /is@؉z6Er19JR+CXJhr]wӬN3¤=pm;y~AH.xXM RgL %QNE+lW,Enڧ@q*%&\MWb ;9U?zT=ՎJp5E\crI|Cj X ~Ą+8a.3>%:UrjpONկH`rɩLvS\*ix<9Nɩo b%q-TkG6n=L5HڱU^Ko# VZE+ke,bބ+RL wtD"JhprƂ+RWR넫 J!Xsb4 ECɺ"4*&"\`JƂ+V+L&\M W}ͥ'LRQݾ|xAIMxrS{T06ml09UGDce\c2H!`&͐MpBhSDgș* XUU4 Ѭ+`.xUv%e:>SFXH0W,D3jW+@W$hprW֘q*M P1+ *~(\\uEjX?4 WoWWs}ϥ`kFOXcWTv+W*تw>"\`2\\ck _?h&+i*|\`1\\bE:He7Y}tp:&g$R@,bJ+Vi+m5 +W,Wc,b^+RiJ"xec x+y`d4"F+RP4q 8\>)3΀9 pj>)u> 57`롌7+In%4xsɩJ-)<_8?ugڟuP~H={T#Y~Ro=< }ٞ;m_^^Q e}W/'=>ꆚ@gZ -%Xk_۞D_]{k΍(^1Hmoku!n/٪4 5MM(=} G_Zo?#BZn};jorydeT bQ_7runeC)Bc+Եo GFyz0* Nuu_e|DzJʒy%ˋed tDeQ݁H]nǖχV/a0$y9.")R)_ȿϪ&ER[>o$U/kuڭ~kHovV㩉 _#2V,`M*5cNF51HqSZ+7j-6RG.˛ ՘/fmYiR oj`:IY&Z&-F$\KSGC&TgXƦf`fjf :#M1EP䔄ջWGK8F}Ց!:f%Uz!mz{&UB%%snX*d̐0r!,4T[ |2aW"PaB$b&<0iFe~C۫e*tw"Tj'=u,s.lbq|}1>sys٣H/(7TM%QJ9A(%%ɊM sHr-ʜ%x&<4}d SmE);I&&(( E?aBiE1k#U>).d5#? j/ <$apZ[r^*QT3j,I:'Eէ (ф#ST:8RӾ""7%}^$\ @L3+l %D: x{" |%TeCݎ0dɣ^f D Uck\4ȔSk|FmxJ,v (*$` J.YjRsGX) @&LWˮė$Ecʳǚڈ&tt%@ ٠#.LA45Z*3\!G˚B`^!LO5VX4:TZ[΢?71TUPzD$Oҩ?$J3EROil AA\J= dHS᷒2UW@lTB15V^S@B]Q[pYV a 1zkR ؾJ!e#fk9clQȌ:Bi*< 許Y7ia;"DDb0x1Ő^@gSO` bXS9 i+XCeұI \\'+TP@S1 Dt0GQF5ͤ3kGRP6Z'o&z\F)^eWcܓ.՗TKOY˨`)(zEkIk (DpP(46j"| $y(*@ޣƻ ~,B D>84iȀgm| u~9X0~)Y3IuƘE$%rBDE}3*ya lz/kբKğ\ v VޅL|6-Dd-DN `RP8xoԉ~rd :JrJ2TPPj7z/3~=4D. iՠ(}ڄ W8n>Ly 1}Y%@ V>(28`GOx>cHpQ E* 2#m lb˨8Tx*?z]@,$!Z*=ʈ~6F2= RbRnB%| \$ղnR"T2{{(%'6ܖS5Y$Dǎn:L>^B ;f-D;)f brtJ߲ܦ,+:\יIi`cmV$FOBnck`&.~)8w%[Y:Fm ӚZSR@ΓDjak7vSΗ 3_ 0#b_wCP^Rt CۚbCѻ糟֛-t^mΤ|o}[f~rszthN[O_n!mM8u {?&5nse~\s#t5 *&)ӽZ[+J $X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@~@Fk^Ȉ(Q\b֭5mJY > @.lGbX&j^|jk ?oV5&j;f XoK ƾY/l7\nfcZQ߸?^=]tw`,qbKX'8ĉ%N,qbKX'8ĉ%N,qbKX'8ĉ%N,qbKX'8ĉ%N,qbKX'8ĉ%N,qbKX'8%NlG<>j]٧s/ZÞ͇U c6ۼY\]F۽w^EJ"D? rN4Dsd8o'd_O+DA9 WAt~K]Wu떁oO}SyqZZEUU*=cI. ~Nw0؄#νf( }qz}@?aݪneyTgc7v>풶t7za[r.ƹσO47@YQoRnT:AﱺY.Puǘ1ʺЇZH2} ™YRooqEV=iS? 7;ݑלz1_{h߷':oE)i6Km$Ho5As83s4d؜}]]<W|8Io6J8#n]n_?|fP9пoL^\>Oi{q8Ǖ[(i"lPY?۷uq}}_u8+"‘J@9墪S~&&V_C2Ez[jXb29yNH((]}Gԯsˬ$QdZSBT,5ec`(9$guz^V>9ǘp\SQwI%%j* j*H!E.{Mc(Ʈ5#ok;ul'&҅׀):}u6'$w/Q$p{h A{*SpܴVq*nZMi7UܴVq*nZMi7UܴVq*nZMi7UܴVq*nZMi7UܴVq*nZMi7UܴVq*nZMiӴN(oN"|9`/17rc1Cb@E.+EٜZ]Py3Klolj&h?7{[>+SWc*j.۟}.>r>kx5e}C4\.h5⸮} ^{Mz6iLُcK[_WoVw?Qhw\ ~P>ןǏ$ ;o lCvsRAkBHbwfVOt`[ӷz.w$GNdڻUGRWqSIjW,fq5Y\jW,fq5Y\jW,fq5Y\jW,fq5Y\jW,fq5Y\jW,fq#tk?̫g~]'œ q ‘jNb0 RZJZR k/oxlc91n~6l& Mwum_m#UyH.U*ټ$qm @"eMp"%IQH#6`ݍ_WQ9BI TIT_oӉچeKj uzBQr}_W)mQ?-UokX2 \S 2(re+Vz>PP')㳠{.<(S@[܎VňD2ZuQA-uTPP[PUv):qZMT0JDv؄&|A6vlJV*2Pը ۬+)?jԍͪ$'Ey>Ywy:l){Yڛ'pU;ʟvr|4AXzTk;5LR܍Eo 7,g'1f,J^P3#gf,gӅ8P]( B£%E2LbνKUNO;1hֽUd6.$>vv=+cE̊åMQYٔyY O!;{.0½geɕ\/tBߎQX6hSXc#gqXq1EkC*U{ vk.)%B!F )"Z\I5kV hՇI*ILj!9wG&DEt#(!\i}t>,Fn}X;8C̊\4b18T#Q׈8F-p6 8x6xhM4BrR APU ǓD-{AXuWQ/ΪF)YKՋ^4^V(8O"hX KQ]d("A3Gn) >zzl\(^–YaZ<{h8n" T#3k+}FQöqRViVFC7fȏwv\%Z<O&.5/CKݾ]A 3x^]C{{S:QLP >hj5Hn4W̢HF8Fȼܰqzp*X'$chS+@kz) eSsJ4([1}ج;Iyx5CߍfhnQMtgu>Mnj`C:gPx} gMAzkSFaZo3^H#mFyсK9gLסW=<MX(#zD=!aDG*I:2\3؊$`ޡ"pk"'Fﯞe> bgV57lOF$Rke< AmF[K 3toZ==?{uӢCf?JEԋҞ&J< p3xM&?wIڤHbH11% ,nd D['jYW/WzOkitԴ-'>Àըmƫ*]5d0V4nԗZNOzI<C?|M i~|oA&dIfݏ5wŽkho&:wG1-?0Pxw7b-U E^pVયB5e/+)U۫ym-,!q5k1Ycjs& Zdh?V)'*}ΘPgm_Ͽq5]Fn4mI ipL;M*"|s桯]N,886瓻̵ڜlj{Ѻ8_2.[?,H])Ͱ}Qe"/]rX+eO9v󮔚5}x~qS9U٦চ4 t4M?#Bo`]7O#sݗExh-Q@H׫_ߌ'L߻-||sNkF!ޓ [cş|Z^&޶-Yz%~`&^l;|oc*k4H/=K{瞩g*5\\27 J'k\G T@JP+碮Јblqzr3 ":BT&#`-2Rrn 4dP-Г-ؽGk>@{^3$3v޴B*w?npI-aa:zkB3j h`!HkSf4++"bEwVEKQp&Z 8|>`&EGfL*)9%xx:SǑw+_'}-26~E;Q L2{Č0tWc\!̖IaMv`JhH*ʺ0R&uιHo΂N<֟xij>gh_G?lkIf7:)1| Cwٴ fK!mlJ`+A|<)'x ޜc}Lc>43Guz h69ʉRL4O\yګ>Q 1PcrrEBD Dy %,wh)*+FΞ1kooP3ۏ6cfrJ5SWݹl)1 $CZmDjqvj1 ~? gڮNmy8X;Ww25|Oi7ӯg3(ܑ3QW:;X}vm[m>0S]onS4=7?on"xFn2E>.lݩ%+|UrPD???S$ h>DC1 e %x^\L@eQx[.pr*)}GIPDRBH`(jc $L+Mx*T9jjMC,lPm\gUNikdz Kd)s71Ocbm `?AtFn+-l &TlEQѧʠVo"#TJؑ Ѓh)oWh",ROܙB$XƘg2yKZ4y_8:օ-S7Jf*A}߰0# F>S)j@ԀNF :HblACe~N kt?C˱j ?QS{wDv]O2N=Vbnڍ9յ(&*Ff+k\ӕ"TDe e{GTFO\U]d]OVm~r yE-U%ܪJvh ^bKq{.C6:t{ѱeJlTEnA *x ΄|Ysv>s:3Sf.̩\){G`~3+mg 1SIu>!#.68\gHF*D RcK!E#gn@gtnN }gbtD:B`@ArAT^YGj,Fd(!g(eJKäHQ.4.Xg<2q)XDSbXj30Wv̔ -N)BWS{yzͷR{/ힺZft,R$ǣcMJzj$^%E Q9Q 08L1cQ&Cp+ƛ:~<^~mk-"ZBS 'B5RYPςcPȸWqV^{8:/aֽY=|o0BR\m$Ĉ&@C ()r)Co5#Hhh'P'Y 6XDMOxQ4רN9(zA*+Hti3Jg[cԦĒFA-PnΤܳOvn7yuv yq>2,F9Y /q/[PlC<%̷<9vo&s5~ # A uGa@`mǡF$,o51a(.;rqÅQ^đ+i#Y/stj_C2&Ր|j c$pu}+$˵u-g@0XUiC(XK ˻1U;AL+(}X,ӹgv зR~|)DN7Wxt_Hc 0LpZI&(D-Wp} F&/1KgU "?'glIg4ܘ"v;z6s‘oG trxk%ƗtS oFf9lZ^y> YLxM&z8ks@UV:d]ucTXI&{$5aT0dcW) )~\)C+*&*uWٻ秗KL.? ( b.U8ޢ3M5 MӴhr _]k}}Ԙe%[n J/O 3tɞwOfKm nM`ĝq]O݆m~SC:̍({^IgkQ") sz ܗbO:khM<#W s$' ExjIRCB[,wJK1(a! 26jˑ-lyy{Q" n u:HPBך"q G-7(2zYq+(dĭN0n[.Z[ rΙ`?CЁŶf Z;NK`Uke2VvMg/4:D9\!r \`D$46 uNӍ,6ih+#G| Ft K|hDBN#)q8&ZFLecܠsīyz8?bԃ~Vf~?Cr}"`UpQhe@Iض\9Łþ3ICIQȶ;Fo?hJTM"TSO&Z@ nTKmo^HENǝ(an0/iYW /{DD)3ln%9@FKpȲ6lF VX͊5jhk\GlŁ\aMtqk(h(,yC?< !mh7Y;χ2kdY+e}"! }A җ_-y!装&qSye18a).Y54 yo3Ϣ'6rWZ eS×*d(חW YN e߮B\P=8ʲo$ |!{ջ.zߊg[z{^hLjv_A2d"{{B7w.X!=l[ I6b;čTFuPag":3dcne9*FlCpjG 8r 8rp 8r@ 8r 8*f`E-&)ҖK jJHe4)emK`B<UX-zgMIDk1h`a ƺ8)ȦN&]h&DB-A=mxU=z[1*_xFSIn8 7y쓑":~_OT=ׯu¬Nڝ\ٵa|y$ TA7 AJOa:N<7cWfY!̡d>^Wjz<+tc5ϕ\ފ&OWkgc^g^䕙l6ykKpl t\0ws;6yJ,BV;=yh‡娰JҠ"3mFyI9&$'V8'{BqƍPˆN(#B@r}0z)#"b3+Cʘt X'u'8,Z1e;5t(2ػ6))pD#8XsKS: MB+D@0Ⰺeԁ"Vm(FBMQ8y>X9^kOxݺ?5;O^.r[Z$wekrW8[qʷ+ɻ\c}b 0h\krk2W K^MLuGLUid|)* YJ[O]>//a3v(]Uͫ$qZ]#Wb?ߌ~o6VR& p`%BPuT9 *1j 3ꄷXOW|LCj]mKaVQ; C2n6Ja~Ҩv%)Zn9#J+}{uB!+6n=> EnH;PYCm7 d͸w"*(v2BZUqBXvI."B$)i6Ű&QdT?pj"r"`. $9VD&GAgNA1AR|cx4m}g-W ɼhFs1umֵ?,us`qe^*-=!*r}Wxabh,]~HYРEaÉ)ꝗFj).HltzehuHnM2FsI Micc$a8`H1tE!rP"[ӸtVO5o)z|pz;p6̖tFs>̍)b[gs;'Ɋ^vTJ`M' m-5CѲrևWO0i-+wtF+'mխNYjX9h6aI sX>.U}; Y/f8G8URWЊJ:E8s,_8|߿pyuyp30 .غ n;iSMC{b4-xiW59vm0[َ^"x:C z7x$E֬r ;9@aTq;̍({^qWƗG"I1#TDᾬβ}@ċ\_GhY l#9a.›`TK"[bSZjFa A`Q[oas=選6 A^̓^k8S+PDg!ǭ(:Ft:0o|ke+ghlM,R:gȰBY6F*k w :,U /MyV"I>塘W6c*[3&2QCLv9,zb#{;;raiFFf56SpQ`gyAF0S%0SYau4D{ Jqb1waJ(F[ :aFhNPƆ`cY#Ou.Rv #NWmn#7OɝKU^v+$)X&l_c82 J5r٦f4tz}j,$dzBy çۆљ̵J d #s#W^=[f齸~x% $!مsdFXS&9}RL%ˠ@2i)6 " 1@ żBU)^"p-<#jw"ˤ2C^' 8FrLZ8RwfA+.LQvg5q6۝Ĕ9^(}TrW61BJ("!O6b<O@&ZȍYѧԣ oH JL>Ԑ<O7<1و 2A:㜑/Z(|ΉXw,),&敱K6_ SBplC,HkM>;/LT2RʭP ^jlu%]i]J+BcFe6Y.qfF^HiF[~Ϡk/xiX[ɇ7gvDE`&kݱ$RCO,R.)_L"%/)foܟiE8Ǜs3ξPJT//rw.1&$,6(hZUrCqvI*6p_mSX^6ޡpZ9 ihsqϳ ɀO(#+ $|+KbE$CCT2wJwGogip"GHCK7|>Ks3!xL-[uKV ll~~>WLiMD|Q7ƻ\L%_r6P•[[k|&PYik|![? ńnBⲁ`FZDVE{jOÙ9}=AJ 0 TԢyȨ$yŭD2x"W}UnR)ie6XR }%O"qYd.s>'! wv~U^MGnX M0_~<=]RiαhA S8(tP%2GqˡSLId*jQEpM&ϊO1yL`T/cʊZM͊Nm"irwׇBZmlEJVIr [uūE;÷>oodwnÕ&h2+G?ݪ aV*3V`gši5.ܰ$;+EnF?>"~|G֧o}6K1 }B#7>i|c<]&8M@'"6>e=n%~9\&soɾcyjA/5Yd [ԥCr&/J@0uwS )KnCL>$pZx[L#g*{ yd8xl.ǣ/tHr;k{a}YтyTߐV9Nly#|.>>>b:@VL$;,蕋&h35R GN#w^zDe.wӸ'`p>RSRAC."Y <(06dt! y% !FΊ<:[8Æ&dfU=|;t$WHDgOShGX=Up.+' rD>H A8Og#Ϸ|ո &-Ywuk?I\|-=e'Hg_&ip>m&e9^w:U/BsH)'[D~1~]M!]"W}?Bs!VV;3L;e'5_R} e?hzu8𚊽zB6 "yZ\Yu_+·0E Lg{s]HԹuK7_ 'yt>.:U{wY.jVŨM|vW,d6M,z͍nqw_=wnY,9E.ȶ ]͗|$vUF/Y;PZgQʖF*o0 2Jp00uvJd)]gSzr`7ۋ kaI"mu VVyD\@:(Zf?G~T^6OL븡fs5mcaŴ[Jrq쭣籎ȹj-EGz9Q4V2hkN}ZJR9A}$N sMP5YB%I*+'HP'MlB0zeLO21NɌтc\+52ǾWθojSlŒҽ}]ޗon?]-[xX{SL=a4Ň=voo>} AEdF,DܚdZ{x=!=wd&!SL L W|Ѡ;ӧ\3 #FɸN 9è`&y1 1eAjҷ#6x\;OMNxscP'j0&f ~x<M=sG_:f>}j,ɳ]x]ҙg&f~*˛YiNwɬɻ~׫L]qN|| ӈϥ|8&0Oؿ?ݤ=K;MJ+MI/ZtKnf.=k[,\wۙv@N2C"\ ɷHЮm_nuٮ)Ӭj=Mni9laNxfokmno0l3'3}K7|ӏT_۴A&+w$Dy=8]oRǣOAgqɿzrGWWE\Z:\)-F\ UWc"WEJzzpFpCGiR2֣KDi Y3Nv?ae9~Xj_SG-3u0 ikq.2ݖOd'vO0=h2%g ?4tSZD'pB>Lhajd 4Es9j>.}]2\2N} &C, Hm^ &//.RhݞwߨRoE;S:%m@$XHłQ01ce8 h,h@8AkW4ΫYp~ֺYȻw5RЀwq>p@QWsD> 9RĕGi ;tHz%(V&aGWE\kHZ=]Rp*'1YWE`V \C"p=\@)|ivѦv)σK?iI'[OJٻ"N['oVOd13'IDd&>inwrı<,aa94Ѭ?v?Yw??~yOfnH<`)Kɮ4w|.o; z.0CUIwH;#Y^l_Ϯ=c}XD]!1~"$B3TU%lۛ+9]fb'OZ*,Eòܯ.ǟ”xjݧ0>mF{x xu_֗,C9@HNMq;.1k>&Ud4xsݖD! 7l"Mhu-ܜN"`A^ȕ/&VJOT ğ{Aq{PAq{P>{m)F5Aq{+Aq{PAq{Pc5M=~%B=(nۃ=(nۃ=(nۃݳ,֜uYdOrޓ܎'O3mz Jقs-/~]eko~芃ooyiuHߋ'<p9[G >ZYKw |: +\0u[ >k{Awpi|)ٹdem< ˃Zk*̫}g^|Tz V,GeA+HքsKIhlmoi=Lmrs g=<!sK,qLɨ6"zsO3ӌ4FFZRbFbk&)ҖK j֕ h`I+:MR!D!e!x0k5f,`ZFL&Z3OHKDI fxjTws.Ϸ_>_\^H;z|sA ~>wǑB{=ZLUG{h_TC~^?@CF]]4varagQұÉ}8'i8 v8/]>(N`BO;˜Qc94H9a)7j<#"( A(\$JB4Ép84t>_ɽpa#z،m]vmbwwҽꨰJҠ"3mFyI9&$vV8'{BqIGia3j>Z΢?6y).o;D@Ćlh3`MiZ1%S!\,6hqXHmfh և Q"! #K9 Q uEvFB<7XymսaN1cSV9iSa`wQhS{,*,+Hx`rgn|y95Xg2W*椦dJe7 >ބZh|oj>\}i4ZhF`#뿉woBYr%,I%3j}e gc[pD~>+0H[^OQ=1x "O٩< 棒[Q-Jze|<-[qPKaq:+݋$O'WjĴ'ۿpyH̹Ⱦaw 6S7ٻ`uF߇/WyY[$h;%e.sAaĶqˤ>W \8PaP--g$^ieڻ"Xꘪ}`mHv#Xx"XVlèhKveܻTUP8e8ra6LQ#㄂Jv)."cQ8Qѕ؄Ay$,ViǬQFYJy>%5KȳY36զ{T+O,Wu(pڵ?7K$_>ĊᮛZ-eGsδub`.z(b!hh0Ips9歈L2.p4K٩By@`+gx>z^ay} E RrȲ\̰4N+H$#b{Acd(`ܲziW#a-k?g9{-b;FJK|&F z%&SUw5#H`РEaÉ)ꝗFj).Hltzs (|/H R̪w%% K4QBDA!@pЈ JtJt |tͣRXwīŕqoaZLc fNc>~8j %\K )( sٹyx!cة.SB + ~uU(a˫Ew)BG q^RP;4E.< CϚ@0! H ˥CRjLՔ\ 6pguQճAaTM8v ^F.mveJ%R(_۷ͿfM.i3Z`ȦSJ"tj }ƌ*#Lfss[gUt,CTP>x{39?85̖9#|pZ[v&8ƣn&% tqa-Cm UoH/m>Luh0ևѸOi.> o.\&+&'hգ6lmjt7"mFFk)B6YRGgѿwЊN/UgQ8ss`ǯ@;wo?} uvŽ$Jx2 zUo'?=bh0^34UlE|qUSn0tҙ[ O7ߎ¥̆ۓU+O{!ݮywX+@6eu5(2WwQRE$w.}u8HND1#Aynem@yS .Jhޚx_@#9a.›`TK8E:|P!q3* cAP{iF|ǀqlxuHx!4^z)GLprCR+瞅h"HRJFܟtrSٷ:qxb.΀֎2Й%P^+?jspc{Ξ1N[&U}ⳤ,b`^OO_>v"g4`+$T8-%& })H3tMzRϓg}2Ft KA$;@4 g5FSr!0_E* ~%l5wxM_0U K5e\|ỻGxUpQhe@I[O΃\v$R͡%(d{Dxk5ճ)P-f؟Z>XbΊ}ľgoRnͿiGI"7CXip;jh 5`2ZbC"{5x5Y1a`q .ފz0H&:8Ek(hQ[ ӶB0ouJ+'WkկY/`_Ƶf,`o<Ñ#J10o+^gLte"P5ˣ^F!OqrYFJol w=LCfƆrJ.tZ쌴9oc9T 9[H[4 UQ`b1wa{PƽuH ÌМ)X}[i |fNuUaPcny-{XN@U7`#W^kjDS 'ZCORp ppL][.HE&zeHj,<(@{Uq OR>g7Rpg0sO3Զ>~j<{[PA"+R *%TCtRZ1Ps[a/7} 0 | vunKuԴ[gx{hDϗ|xL`ɻ"录Zm/PŎQa3NGϭ֑jBe\q-,R8RJ.]{YK4+nzr{Ӫzz4xK0  >u jd.&_ߴ)C47Yl3<&󫦽ʴ'PFn[ƔU-c]Db;^n,N%wRt<E<=syi-Lq6{jS/Ҷ7ʼUЊɭvoR1|p3e>{WCoӻ|lФ`oJlgonz&S^pL|yߒ>\'o0X֑?_??ԏPet9w:h9HV3Rҿ[Wy+SR[q ;(4ļ#c@ݍEq}ŵߝ.M5{ҥ\M.[yMRhYz Z\]W;~܌)⎅s1i{!'~h]`{gNhZ*mwOOD`x,nnf^BF26*Fh\jh6țK`@KrU.襰lVc8~%A!$\pƔDR*X5yEMY6HjIg&XWY)!2NP-"jMjpUkSWz3OPYADFC0)Mf})\RfXw/钦Nۑ&TȈA2p/'xDLqAL,mp*mof5†Ĝcv£LXٖ!;HќȁP -'ʽfz; l%KN(Ѽ"ړ@ertY.qxױ$-еW\{>X[p iY/^0]m S,Vk"O50W ^{dD$(NV(tz.oEQ粔TD~Pݵƒk(D➌ODS>e61J)JЧ୍a1D=7x$zCWAڇ6|QInO~|N+cSTz66MC?_F_FцׂGy74fl0&ŕ6}8 +̵@gҟomעɩ7^ȤLJ]ݤ}sZ e<1`96m[;R־@˚! t4” [2Bż(d<fGQ R2: d۔W\XҹKcEf1Z\]JG }^{gݓݡ,v %c$"`'Pr\פ,MBpkR U5)ܐd(SLN"L )^zV]QŤMfik _v%,;z˭[;9ӥhoqSPwXĨ0`@C3E:9JHޢDA>cr{$1K.zR(m!fEt)bA"H63VCY5[V}PUBskvϮea#d_wbvf?qfxQgvR2y2[$&Fad*]佉tɉ`Qʺ-QfE8{. &RڔQ(c#d,MʚRf2cWx\̮v5ؗMe6=kAlo Q 5ڨ˄2.l*{ {͂ 츪|j~#PؖR5x'\VfF?R~|GL*qQSyPm~2ȳצ/iͼXL3K:?xr{`4&}d\Oap|ah=.FPZn^LJttYk܃}{|_" 2@1C}r-%ҹ4tq'f:_z-oݾZnXv9X=[#f0ǖh?;h:'s~\пhz/s5-;4nדNo˓7=Q<3dn޳n6X1iY_aV-"" jTJ"~1y0#ѧ }qz¶ӠT*tC׼L\^]џ[[,4O>ǭI&qmBpXꄬ؊2.gTֈ[#[0ؘԩUA{tUPZ+bBVY!+pO ZyP뭫HWhSr 0 9dt~~5ҕ-ōi8(.ڊkTLXR!W8}٧o}5boeht"Iw ^T@~'sk%S| cJ 7ǶEn6;9c$Y/LWσeYh /CWC[teAW}'DW1&O :DW-ʮk7=]*eǹ>!*A ]V ]ZštUPJ++ qBtE'CW4BWKm|J0=]B2é\ 8;WpaWO?繭D{B4M8'cUu J[fъkkM.ôkg'P ޡg~湴~![vIIZ T*6%_ 'hg`{tI^.͎"+s,E+k7#K콱ZFͬRF9kb&ZcJbjN긹]'Wh)d9[zKFH .5 ͭ7ט%v1%3jLҘrO,mnʺW > < ͥ<0dB5YTǜAZ|5LZ8 78 zey"^gKQW١zZ]v`;"sEϯ|.H.=iAx-Oo/?'JrST{q] 3V{ @H;JEZ&;rIG#݁Unƌ& ]J(Ш):RdbphI+J58;钦+הHJ$-#iTr\XD*22b3˽ n[26$4K d'KERؐӀ>p.p BtIssي͉ު@B^c)DX2Ea̎楔 tО<&( r3ƃLuH&)&Mv&靭G?粔F4us卖 y-/ c=,X# 'hM;O} hFCqMsI8~ɦB^w.> sZ[ al ݧp\?n>`Fו?`p[>7 |l~}oFz <8/gwmni̺ٸa>Љ6x3Fhej8ci5vJ{L*)ep:[! \O% "|~A)D @`Dup~C .RYFyTc5hz.Vtߐ .w%eO,O-N4AC"h),$7Q y )rեrHɥdt#)E18-W*qY29rH^K$ewfa%p{ ه=dyFҦ A0k6pU`i:[],&If4 ېtFI5/br"aR alWtQܨfB^tGNC̊P4b1U#Q׈8F-p6 rvV9 '/ MVHNJ! xRT#ZC Am(6)MyЂg^#j J;Mda,Y#~:=i !:qɮz֋׋^\kITDxMNif!o#*2HiDD8L@[J^܇^<˹w#>JvG=yTsMLg#7z_1&-n(p-` ֖Ya<<݇(W, ,ѼS[DpP]uV7fPUVH\ɴMK^VSՆ2 To/PC?9QWTx.,!S4xgQ7X¼+"J/s Ӫ>9bxcB@ Ņ^q;RqB_ ^hN^>㕉'w/TҠ:I[sq r{]Ntۖ91$|L0s[߼rs #а:mb:;62#!mu.i$SN)B%V^z[%&sIu4=Jz JRC}mVEV2.ʁ-WEҾ_h.*T>DZ)C!{D()Q<a U[k4Oa%D{l r uןpmt8a,3bN: ?\>]GȪW߼]ƟF+SW#0MSC)Zw9SPl k2>7d2Gm U6@v;|^iB +_Jj{s=x9UV)2g0V jt\ ZC 1^_ j>RIHwB:x"UHdB`nUFJ΍ă\NU li_  dd m>ntuƇVh٭oInVgfak:S 5r4yVLG)XbPfDe+EwVEKQp&Z 8|>`&EGfL*)=9%Ax2 W+_zͭD)"޲|!˒ $s˲M9/ Ĭ+0m郦AFk2꿿]5C}#^M~p);|y(PN'IJ&q}?OVRFַސ8C)S$F:e)Y.37ZKN֮ /mzGWFCmc)key߹ޭ[Z7tu2Tm;h&?i|y4>z260,Is:Hr; QG6z2A<_W{; t;9LAapBO,`^ m85+BrKseFv^soAiNU[T;469ʉRLDW'unU Q 1PcrrFBD Dy %,wL>(FΖGm37p8MŠVv3^$]}ٲxڽiKOz=gƏ6av~r o\KZS;SW] 6)0#$Ԕ[WGWU+r5ٸӮ[{3.Qiyaz~ȧnzBSDN;EJ B zk3Ũ ׎PSePb+E7}ٞ(U @<4ӊ DKy \yyHK>qgR$ Q`cZZ4yssΨ~^]yg]ՅzﶛՃyU<73cDFtJ2EȚhRֈ`5A\7hHlu?tuZT련R"xۮ/djBwm NDK59P12[YMpMWFPqxoRJ?/e$ˢ4cd7d_fk_˰ٛu]]UB-$۸-//mAZVl >xυrhFf/L)͎*hqít^]1HAoP$oLmc#ބy!(iXq,#h,j*5Hm4rV9H%bU6k\ Xmޛn jGDBQeَ+L۝Ƹ'\q :*Vd7 $K,C> W0e 5u偣ZS#Eиp{cAxg˨Bdj|ζݕXX&Ph֒'ڠ[*H hJ5JDtHzT\f Bm>:+5C!Qlua }ia݊ՃV #$E@B_]d!vA{~Aϵ5m#HJx4F,pi,g &'<( 5r<^ .mRGsp)$(*?HI( ) m[6>^_+flg^^9ߜ 3l~~\i}R'J~|C矚t%\L't8cN?37 &/Q['~%r$!.?tχ~ 8hjQS3E 5hd a(.+rtÑN~9\Ya^枇 >Nfr apu,KNbkղld9"Xh`(>}7D+\l$Ԟhד 9׭ϧ3rGf9oc }SW3&R^o˅ ߼9+ J)٧o js4O:~sx}QynLʜ;ٴV&?&O>TNnƗNb@1}[^rkd锃qeu>E/n/o!u#)Y;G ];2|D T`^|lG1xlu6ɺQ*9\G٨g2> }f;1=ONЩBkY<g܎ߟ>xrp~9eߧ?38lx>{Cz#M ͆Fl3jrf\kƽ'>P6Rf~~VY'exx{:s#_,NMԈp 0$ż?u_R9,xCv/ F}T/ AxlxojEn_lԏVģt$ *.b++ ߄࠙AaJQGހ+rVĨb:VGڷa|;DQq>N( JAN$rq={:BHwsBW΀/&ri+kL!LR5wN/dVD2I“"A**D*PU@:CdװR9[ݿjW1B{YN9s6c{ _.{<[=hqK:cb48dcQ1p.@I2EC"T@ᤃ"1ɜhZvU䩥/3S]4y43PFJr-a?BX;"D(HBU 2> ^s\t߶}^^r/T& |egτ NgqO_U9RDUg)~X՟I{͙lc}|=v2by#>k9K/]LX7]e BzPn̝(=*!}k)2_1Vjtam Z<؁{aGK*$Me2Ios#q^{!Km@Od#هS"OFq |(rf28{ήG#4sBMŒ,4|1Yd- hJl&tKr3\;JC!KAi.$<!δJxT =z@Q%0JeW [غn]LZo=^m^ͪClY`bʭwiyhpS=/9(9o2=+9Zw .ҬW/׳l/oʗ8KhWo_Z}T..q5V7!\j ߂r4nE*@[6lMS ^Ta .lGmkmy~Niz>h]cWqfb6pBh  d > I)۲]X١;Pm@";^~}K{֋Xࣅ/1<ĜXGrW|kVY1I M!B`swP̼搡m=!(;{bғ5I Q0"li81dy IIG{Z>J 6S̀e8<wYEap?*2%O^htUPaE2MFI (-AXy*WeSϢ{%º#G% .VHb&@C ()m)"?-}:$D%<#J3&%xQkT'e-ʜ 5+HuޛޡԜQ.XcRbIPP#h+=@IQiY)gR)t A68믲6yu4>M837d^QBc\xWZSBƞ0MG:üΥ-ɫؙqrra)#A * ю3qVe|F3QRTA0Nѐ]h9ȎFzG1n؃bl(@Jl$Ԟhד 9׭ϧ3Ų*r6_OǎRYmghMjQߖ ys<WjSL3!r#U+r#5d>ωK<>FM ʜ;ٴV&?&O>TNnƗNb@1}[^rkd锃qp7_^, BFRv$!׏ԯvefBb9'\BcvɩIQ{Cz#M ͆Fl3jrf\kƽ'>P6Rf~~VY'ext#_,NMԈp 0$ż? <:~EE/wB ;*z|P]$.} XBبGHT\~:WV A3Q)=+E^z KYPP[iƆٖWT {p @m|;'*(28&ʝI˙1$\Adt*Ĭ1F'4r33ٷ6vxb@2?{FF~ڽs;|_mv ؝_Ag"y$9 b?dٖdYn٭$)v5Y+oA;,77:x췳V n;O㶓\딗kfřH]% 9(g0aH4h=ȉoֽc40YF9ũhX$i .&Մyg(!3Cn.?Li^;ѦXckSU\.sxibD=2>I{j+x5j<5Na%G7(o'۰Awe6Ԇ0E/%̳wbZoiJVp0]|9X3/mZOyVMP<"K5/W:X1'=q0ߪ1gF֊)O,<%QI5όhh&㞝B}bH{Mn|/^qd^GfK9ѥ?{<t-[Ip>OD}@xԧŲv˜}x[U\dLkCb#֕ ye;|wI~CPJt>N--V9C| BXZpBREnd YV#7.P፧:|qq,<҂F0c*ؙ8;תqO>Lq}n8=rYnSd>&oɝ#kusG%1_}*bGY ֔Mf,.c,Y,KK18#8 9ZR ڼ+q cDp:U,5 E? ?h>quJ;VĹ[QYIVt҇3N +]S*R?Q w;ud-T8̀0χd)9^Ϯ$FspwJY(J!$' 'fB2N AR-R.`R]#cg܍qΰ/BX eH[\ffcR׹_.Pxy8=[f%GrR1FΊZJxv zgojr K0)\"C6uΜmOs>ưhU> &"m4 A) AP` D8.DL&ʌdXGhk1[ Y_^̦_VRWv٫і~(gy3lekkEeG1xHd1"RSk<#-D+L1Ek%3h4`ĀA sve5s9g"D | dn8{M= ZO0 D(%cx֥䕤!H\Xw~5:ؠ.C&Cw u/RԵe7rw sCȎR )t Qsw]+m'7)J H\pCJʌNrMK$II$H9!ϴG̸ͩ7h4]i[v]hsOV}U(;LJSHI1u~2Toߊer\'zf$5lv z.C9QvTM.5?κ5⾖zo58oYp۝=GWkkbk w[{[_a SX*\c&o0Ltmƾ꼗{]o_sϣUP߶W>;"з)9eT)cI7ԇl_IVM.ON\IhW(o]D@͙i{^6Ѻ$׺Y0 %"=qֱfqŕw׎&;^poSC/6,Rs=hRw;N-+2a7mj =ұAl{nWWk],P .f9U b* )SuiN)}yMh"EAˀ^&4{)?ώ:z'" Qr')8AQ)x5c3!o2osz^>/-ll\$zXuđ/nJ9C' xa}* _n^)`R2iL !R+K+ 6 RZp"}~gB%+֞lV}98^T+q%@"8G$BNEh;ʼ9,52vJU=I@' sUȌC}$&zlNG4H6F;CfWtbz++Y7 +cm~DO{Cz,Nj:IuSk*T Hat 6F pPk6V;E!2aRC)u3qvnpIS3הJIk%DK91E#4)P´^Ą$z댷wEB_xl!tC3Qa}$V(pAb"6R ~Θ +Y\&|j]Ϊ#gEBˁnL,C<-p4_ZΦgE3׋?z f \eq9JwRj>3+8"+s4p5X  WYJ`\=CsrW+";mq}e1yV^1~+÷~C4bg^'_^&4E5ڏwDI;쌣ԀbFi{aQ`tȱtVʾtRh\Ip/5Zo'y>M/'j5LMMQ|t Q&TH|>?z1]~f3)'߰ RdvX椘]\o~l2k3Baa~^A KH!)pLL̩G. `j G#z4mf 'Ktg,l g[Z3rDp \ei䜥jgWy6J \eq ?Bi9=Rr=3+#:x,9#V,K$1ձÕޱ)ʧ7%0OSt?q~ZphK{\OJ3wI}\T2AUX裁,"WYZW($\=GX`;)`;&Нl5%$;{oT.ے*۪XIU됼uI2)0UB-{u(9+ΤT axv60jMY/)On`9X夝j䂸 |Aw0N6טȑzҳJ#I tFX Gcf ZoA/]ю`x_l-yƍ\j66Qcv~~q98>Ϊ̺g;c`Bl3ٗhY5:ڲOܩ(E_kV5Mّњ Y;fW֧K)1 4KHbX^gǜϯ^^=L i.fPA"BuYe!ze~ţ5fw_]2Gu*fl߂nCTX^tF} E]V Z!Eh*hmE<Ւ? ;?ę.)AEXr&ϺWom:oPs╍|ͭM̼h׋}7j\Xcp6c-*rW-~.4 &7dAEmPAٸrh..8CIKp=\qB4nV.'BY]h@t~oP7viT-i9YhL,m-CAGB)]1ƴQUʻu֕Q,dP۷CGеy:ɢ!PmQdMQ"MT BYKy^%߫{p#+Y'\Hb _"(\9tO.Rx]پJȵֹ_/nN+T D5ֲ}Cp LMvᛮ_gu.B&VD) !,H2kgV q.2rӬ J]ac&WxbeUy-3+,,Is^l:ۿ̂0V:фZ0 b(CAu!k0σEk݊X1d:U^)4jimCq |謏}I'{H|gU -A :ü&L'ۓRi1lD{5jM\95vߵDb2RuqBtE).t_ -6{X?(G/ז!qn¬'[ky5b U?Á@dсxdT`0q؛6ךݥe=N-U͙Rb6֛L3#D8I@5ZY[46Zflpo aܷ#Rc)mk]!`-$D)SIO6n}v ъB@P+mRW۳•QWZ 徝p+kmմ5 ʽB@W]5v[EW֏\xЪGP Е=նMOztɮt+ػocNWRSv'HWAT kBƴl Qry'HWv+j ]!\BJ;]!JmtJV51TuM51[ZѦ9;Ia홳C5svV Q IҴ4ZvU'X0KjǾV찞hvq櫈QELErERCo{[MÞ+]` ּ!Q HWw n2̉m ]!\fBWV=]!Jsxte6MSRB5oNWRwCWt Ǧ G~!\sWՏݽ)d}*jLH-ZDWXBvnC;]!Jt1#)i]!`up o ]ZEd@WO8mw*wAۿ Xu_`}Bя:e81dR\h8??NvN%UE Յ/ /yk]E`l`̌kcuOCr=ᑊr戋6jk PZ6:XGo?*Ѽa_ܛv^'>xP0ϡP0°UvEn?YT8qvX2oyȒp#$6TL3A#ܭ.Jښ D+~D(lC@] GXFKT}(⚇2<ɳ na#7 ǽ?{hϩóIq_:`ogUN~C%W]&H6^=0nxO{$En^=d-Tv<Ϟ(*fN ߬>QxusZyR]':9.:3-A {CakYVcdXd?mw%]ʨ+Ib8Qv!$ =9uǜ$PXb kmTDQJOuLy`^\n`!tS<pյ&TyoЛ8-.4բçNKbi* #IxZa |0eӗm`yPo~ jؿ ]u;?4Û] ]eqt#zv,ey 7)g˃W MOL/:I N/ڛBv},u*.Mj 5ٍB߯wQ oD%7ڲ\F:h U:ru=!lzj%[[\M![ Z$ v6myտ{Xu. CddTLaA-8+jxIFcD\:A ˯ UNɥd 1PRrR!e96 Բ.UKRZ=Lum3;a0}ܞĢfAibP^ȃ3FJ ɵH$ݨ&U*܆$s᳤)& I&0z ;jcpV;j3dig, G;ҢAɵ/Y4]Jogbpd,z٘l@S/b*Aj|ɥۙMNV,gV䄓ܴgN|'q'dAiqcTa?J7@YpàS(3*s 9 '1Est*@rj*qac-P^Be2>^  y{㻳\N´?Y^ed38W'" <12aW1.yyFc8oṀjpgH\頄Wبscs&lV!$h0c7g5c ֆڍYǶfmu`탂s.WWH.mchnQ "u|E+y*"Dqa.vIqʐ>J' \ %F*pH va>l j>,Ë4bBbFl>eD0##. y8uT |Qm>xGM20d RTOeDk(!0|qS+MyԂ30^ F$<i gl?&hj8fεOsYɶhEs/96Γ>`1G GpTd҈(pڙGn)>x>xqoFMYǖ|Xhgṁ@aU&i:+ȕx >:qs#*Tw_VV`2\#Hя(5C z ߽=mJ]qrO5C@)'Fq̆Yj֜IpNΔt S)r' abh jb ;KnMz1%NWj[[lۗ#wjp" ;()"iyPGxcwǦ 'T#Gf@o\"B"Rke< $H&M:Z|hj'Z Ӄ7ENYK%>hB8Zj#QȮWXlH/5ADd>wN|i|д|k1 rwo\;Lt%TrQ(KPJ~x3Laz }3m*0n)j1A'߯]-Uw['*jbafɄ{񥎋`WpYs9m{*,Dx`zx>\q7(6}9e|y*Ax>)_qdVSQqIEd:zBެ^B%H*7E)'[dhYU1kdK"0kkUnwۇeNOEpnpU ?t~:AR<ޝOOCxC=nwS''^Lݞ˰YKY <A_)޾rу怤Stj˨.cl~sz: ڻo?)N''Wy+? zgpQݺPr=s. )6$]D\(ܱ)aӛ,{4={L1;~^㐦 9tkgJ{ɑ_b:md( 543߶Q c[rIrw7\a;eɮ4P3x"x :fBPӞEhk$jѥkbɖ:W쯥 p3] ۂ{ E)頶%,b^)o[c'4qMa1݆u;ݺZ"Dލ{TMs@捧)m (mLVH2ArbhW _*+KQdN${gz7ɑ2oPGGg q3ff@2*HBE$- y H=R ņ &B/z,zeb>&iz{s,؝͍2S#aZæJ%|kR-}hiP\Σ95WMJO !g2VT^}Ҭ+pݙt]L.vgKJ'ʲ>c>IPUZACHk悳ȂeLII:2,*nɻqPG)TEIg,  [BFp嘒sLs )Etڧ(xP3qvHoc6_JkkcRxt:gN0RFq%2ǵA٫>tl~}\QgK! Fe?`JʞYlco ߦN>}Ⱥoppu"#ůg35C nK>. ߓ^ !KI X#fy^<6Kflxm-#v-z?@ ؽϣOףa/]`<#"ZS^Jqg+hWSDx}^+r؛OeNBW610mb\gGFI`R:X@%5U EdD=~˪H *ț[p3C=XӘ v q6!e7(w@Fm,kK "N `)i@w{VPn}ʻC!;.'ϭB*#:{glqh4nC̟ZX쐖aSޮ-CYEUtb E?5 `Z7ZE_xW]>U#}6SW\G1=K`.qDJ-qzKWX߿?I~' oH^Ԓ&ItY376ޝ5&ۻZ>AƿO.뵑=mU[n}kJy|7sOӳ[zm*u6m;h*?kT;f/ wY|?2ݎk~Aeo@eT# h6FxA8?Oz,w;pR3:W9hML5" ˖{^X8ڊvSyCr$ή;yւ`8sΠB Q2ƓƠ0O;zyк[5Vj|IТKȊQJ`iqa!X#LO:gL+tBwW7܆Te}ik X4#دGo"6ִ!=*{͕*вV#xXrHc+-U"0\ں'qP >@$K .%DYz*sL%O"AjQ2Roǥ_ݬ^u._/xDH8(g$- hjZ$ɀS,#1MfH6h̼۸o: _ڠcc[y2Cߝ7m)\v=j=&llf{~?OK5{4) )+ɒqX-=Yv_aaP}~=U325/f]wjn#cFVlv.m_%Xz-V/Z].'6y2{ɰȘ8',-S琢VM,o.v1^I^t>cjnZjmoȄq'EuWD$g*08iB3I SbNEɔje9 9(lOg|5Y_ap CAlKȤӼ{abQ婙H szR M-ʘu:`̘2Yj2 ΡFCvcL*ȱґ7,B a A˜HMgrԖ䭎I6| Rrgq^Y,G+= MK1dȓZGNZQ%H 9Hz +^LfrݻtLM'e*g(SRV7Bʿ[3c09ʹ%sR_n6j[ .%QϴIY% `*1W嶝R^uW'[_vk{˃Г>dwʹ{Lݞ)竫1.m+_o=<V۫?UvW b!*+*Bp@&YzbH=BfT+!x'&%-$%!fehS)c.8)Š }J\d#eAzyѵ3qvwp:]J8n۾8Qkh251]+wpp}4]WEحg u?얪]xG7Ӻΰ~vrh^䣺mz&'$\0fTuK)zdK2_]ͽYWH%+*Yuj7uW5/ o!PRѮotڢͻ0vVk"sOZp>5g͟rMipQEcOryǞ%Hi_Op &q3'Ho)< O.DsWbm]u~A F_}Vs>Ee4TjURUlJEYԓ˧x\mưjZ fRs6Q.Z3-TdB&]uSD:]5rdC+2g%KܱO3qvmc'>iz}3Sڳ[~ٶŵU]S(K[gtĪ `3ӛ25z^\շ)Ym &~|HݗL*O7PϏ'@`4LOrhm!oExv\S?2{:c"T" ƽQt2bB) &XeI;kf=D.^Uh+&gcR!gŏwy G"0zg M<^kɍ_gI5?Õm{UNΌR7~:t+b?a9-9>aL. $vcPۋRI0Τa*&gRE;2) .vrn\≟Lu;Li>IWPf:ÒW42BK'&h̺z-n~-2pZ{uy[Y?N/׋H]Ulq ~Tq+*NpRY_7͜Hie['>0M(`MmASѯ͓};~a= h櫋OMJXsd0 ~;?wwmY`r=n &$jD*g0}O5)EIRN,SdTս[U]O@m*I%Mň8ucN'`mŧ.zvUŲ[KR/+MuUұ}ZqzV_dZ&6^>.ߜ6fG'= ׯuo~?¾}ͫ?|5"hyuˁ!QWoQhwYu^6Cʵ]VyC1.VQ\ցiů_ofm/k9WQs#uɓt_[xIj)Ċl#A;NߴeG;^h;jbI=ϑT, \O՚n8IM)gg>Q.)i>8|Red:iTN&9'ptAy)eJ*̊ZS8ɜ,/Qig6ѷJwؖjיcKԫJeȪqյ86)/o}N8E<4iALN Ʀ;8ڥG-ytͺˢP&gX:rqxD 3ƥPݩ껑I_&ϧZs'>,wjⓧ~j&~Mu{f]Λ52,x.6ǢQ3 V٨J=8E!_#rN輝8R};/i^n35AGO^޻5{m[Y~s9l^gc rZ:>U;jQqD7Zِg >1# Ƭf2$Y fg5 RO\YĜ}HrW$_"gJp )RNY>VtHR(ω)ƹB\*HE"Ykc!F&K¨38 ,'ɍU y3q~w^Jx:t[s{3ILr\rX=]V]+Jq=АDWl ]Uk*ڵ=/{JW#]!]Yr W F]UNWj))]á ׊3}R>z~v~fs]m։g:c;g't-芏tks͸a++ *Z{(HWCWBK&OW+}vhkfp;k×#]]IR2XUá F]UNQ]#]ah@t@3p ]Uꪢj=+ŔWu1O뫓,;Z[.]mŵ00u!3W_MΧť yιf< ǹqZշwrzޟ4/lz/w@h7\,l*H55973vvy;d&s* E暠kP"Dd;c+C$2* 9WÙ!Wޫp3=TJpp*\+BWnNW#]!]՞7|@te W 9ԱtUe|=+k5v@t G]Un0 h NW8t\*`;+5l03d-Q骢CW]/lsvpz^ ϴ{;g-Jtks+'x+p ]UJ*JFC)#DWUuEWvC։PGC"&`*\ˆBW@k$*J5>ҕbZ]nu7.9;BzHWwoSo#ِpW4] 1h;MWfKV D#t:HaSԕҤ(R3??;./U/瞽^;IJ_qʼQ Nkxf{\=Wl )"$K3h*;A^=b5kџbO.Aƫ9I"ⴙwQf5JY٦ !FJN7ΧԳ}IQGZ[1=[ugQ,\\ lKɲJ$w9[F64(s=Q WZUli0tUEKtUQqHe" j@tU+p ]潟ڽ̀ %]UYYѪ/2(͸j+hKugOngPpa]ɑvz.b@tU[9\]Uuv(Itte-٫d6xmLOnaSܤ/hÿM/ЌK(x͋)Ϻtw?"__:-!͍.x⮩ֶ?}[}H],ro{¨=jsjzqNOGnZ\v8q>"/^/h1UrkK7O3ʜZ:֊^kaGtd*C=-WŤx8E]w3~}~q|~r4F_jqe -`-cN\-4F *ߥ%|QÄٗ Xsdz%uol7|+W[zڋ4,aK<} :Pc쀴K*\*ZK}j@IljćZVN W 6heEٷͬFz"bV 3pi0ꪢ5tUQ:;ҕbڬmпөDCng%gө7tC_`"5p`TeE$(U6iihZY6]>X[}|n@=|5VC{m fy&m+J!h#+V|8*\91V*J;&{AW744┾U+)}qVZ,jG:KU\o兓7/͛QW{}O`>G^ jq?ګe ޻$οut F?8_U"]e줾jgQssBӚ[]}Vmu)J-e36.PIk⶷l5zuG.mκ*c~xiي'L/0˞jחqKT]}c^돖V>,{F\aGep@Y C?4g3/Љ7# *ΑPO^u.j(,7)rϚ_5]{/> .{Aʖ̵+D:8Ғ̃ Ӓ0wl>I+83vX34fQQ; QqY=&]{{,Øɑؾ39m2O-]A5hC#MMI$ /x]aTF[bFGëkJt:+H\yX˸3N AiuQ9ڑOy~vc^ef$jZd<%]JE%[RL$Ἱ^a9"s1Jq%fE:ˊ xhzZwK>Ng9ls%yz8:@/G&)%S]e:ĺ2 AԂ%I`!.R`TY,3\ D@r,F'4.FĒG! Vx'ռ"Mr3xL0VĴO2H4@-/!D :2X)x1)# dyx)VNnkiG`. -  KUrgh 1AkO-4oh;-fa Ǥ ,(y 9)<تAڔ(̂dP&Bô[IcJN٠*.4-"]dH )DKYzP{ X\7e (uk0-", hI;țP+|"M`q_KVbQI2a ( gl0`&d , |d-GE!Y̸E(MW1@ejfL|"Rgd#X.-*z]rRq<b3.cH@XPmo"1]urLh PD* y@6 t#X,xDw`h&.CC\)L uR#K@d V+\C@G] 2roJ6"MEs%Bn,N[y;X)R6W)`!U]Pl Kݩ( E>1i_w/LBf !Ѿk#9B% KQV `"R CNl bZ3[lj`4~AiQ3uEDjd:bV!8Ih1zPT$ U8G"Np?<.`JhÀo4>aƟu2\"uzxQ6c9MOBT^m8@C:*t $]:%-U:Xmꘄj;`6Uu )t]hMHr Rٻ6,W`[@0HfXLd aԐm}O5lS%b7橮[{N7I"&iY!UJP `j*9:,~OA `q)ԭL#&H@@8ڂTBN~d} DU;͊<-&sJ uYI A;ߒ^ntsW{8ݰ(QC^"oZHs1#J -E]r).2TA(< U@%'' HvXYúI!>^6[`E^H"\x>\N.v)M^gT PMIA1XɑۃrmnPC }i3(qV9Z l}AXrdFiC/ڀB\D.ĺčt|g*AkO&3JOQ!,6 NqL*/ Z6ِCL\[յK.ńJjJ*a7uĒN 8q9@B؄r^ZQDS F޸0;􏰫^u#I;ًr+`7,C~YmX# &F4(Jz[$P`_8} 0F{2,`\9 O>a{qRR,7 )]͠A0JNC:[{_Q.hq}LH=Fz1_tX\·\3C4cn+^-~l^ϷFmKƶ2q^}0nŠ-;jLJ#+nD}rY"?N ծ7N u@֡v%'):uEWJAWSϝeU ;]UV0utUQnӡ+7y,]U*\'BW@kUE)Nѕipmo誢G_ %׊Jy]MA`!BWSUEiI]"]iV/m~k?VR7U(7~~g1Yp4GLO_b0.8cL;_0;r(ܙ1;ΜǭZ@Yy8حi3_& tӝA|L'X?zw5F蟯+(͛@ٓfP{,ڿ~6vLQepd|s$U^g^ oGj:V{VKON2A;_\oҙL)7G(QlʍdŸT1ˇs01h?& >]twi,fQa깧0}L>|6<شFIU$D׻C=&x9ك)c"=&bR>IޚMV -kQƇ6|I(e1**ڧn,>6J%f'X2v|b/tU*~tUQ6+Wo{DW ]U*ڧv<*JGS+}PBU{*Zk*Jψ4t9qw=]>o/6R+]iN=`{CWk ]U;]UdWBW{DWl ]U󊞚\OW$uut%p^[B2ٛbGOW@S+U{DW돺R1/tUJqtUQjRWHWyu7ĀJ9JY(gxvzb<{^􆦵BSwO@ʓi}>~YTcs,S1<`6fUF%曨oZQկD=L?Q5RKzm]AWSϵp*g'^p=}Vcr" ЕF #`*\uU>W1J ҕ^{#`To5/tU:{tO :RFrizDW+\*Zu`Ewut9[nc2;v>Ş9W1l$h/8>ww3s׳=ڛtYy^3=#3>>Fhb4GgO^p})̀1PzΩ4;Ҭm{A7YU!DTekMןn |5.]o4XYeXo޼ .~7v\r 2gFyO|)0GcP+*fm+\ٶ Jq*8"XdP/zl:Ӣ~f)._6^qΠ8gq.#(Fq]2/Jsϵ:s✴(YbYK ے׶U$">ѢA @\]k+tQfF Fi.[[D`>d>{w9./۬U}hEAWyӛ60l̀vY3 V`lVpjڤ0׾/ӷnpg|m6\&l[XY7}z!_8bq9g JG@|h\z{pQ~ii\#0%F؍tt¯uwB󳼵%sP >TdJo{ѽ\!ݛyYu\a 6evfݳ{/ݳs׻m㪯}C >IeV0y}uK]#'k!tU1'bvx@U0$TBBՄN]K4k*`I|r_]kiz52Okȃe2Xs5U~@Fȥgo$w,~wMrn^ON&;G+Իo_h=yWUjl~O]s}fe1?JHi=j5&IoM_h`$}?!ny-6Bn_[m9o;dWmv'[RZϤJkh&sx9CYIݘ6F6QDֵ->1HN=\?d/>ߓ^}' ,S_wuFb TVMK棶W|Vd1a6l)[ᘋRrT\ƜB+UhY޶\6/dQ}08u/hڛݿ=Y̮/]R)MFMwk3aOޕ5·T*sI&ṲRu+.hQ,R˿> nZLRJxvЍzY~C+M&S:jg=cYb.{NSxs'vm2StUZO䈻T0lt7/zJWe6WU$mY\b^Wܡ%xK}˧/Ǽ˚nwVFS„~n؆w:-hu22WO[j;-ގa7vL&s3qs͹y/^y`ZsxY%D]ǯv s@(!bvH){V*|*f]ͯ &dUJ$diϵ1ܬ6uE1&B( mb *J_aܣyYCw43άj rt#{x[C炘AX;D˽5Cq/-^îFVxyt|w>#j $_Vɤ-/)sJuFC* |~͗94,9Yڃ˨mNSBURg "TR$3Af\D'jݔ/+PIgP#(WȨhǢcl*0Ek> XDh̗ugeiΖOOn/6UF.f;ÅS#[YDch2鰸` /0] y"}S!38*"!WA2pZ$e$IʨU #>j֝xJ Lɀx(tϽ箺^0螼h}Շi 06g{[g:j ƻYZohKlȞ}EU]ۢ4C2/Ʈ-NpМ*] 7Hűg_ڲ8RdVYs %3X5 _Ov!5ngm0Ca+/}&m%W'$iEEu .d|,B>Jˈ+]D>Jcu &JB[-hdCΔ hf\`U^ॗ.JJQr&v V7%%3S,QSTI#Dp٨ /56h2%#ss7:ms8_,|(ڲ>psFF+KP|;)KE:Ym8_lDWAy/hnd]/;-^}H 5T"D,3eBǤ>ԌJ42uB3f]A+HΎg&]1,U R[AzHZ㴒&tH9Z)%C)d@@zÀE+bYB9h&EAaTnpkL҃yzƿΎ5ӏfk:Yr2*Ós8ƣ/S緼V;_@M-Y%0[:Ԍ4?GӺU޲iamQ/&|v}!oh!ժ=Ֆ; ~(d2ϭ F?-׻&3 A^~b~J|*f^9 ]f|0x%տɲa[m4]I#9`7:8eIQaLҡA\Tl@dHO}ذTyLzÁV Sv9-EdQ'^dK&&YEE+ir3di&-zȫs=jVf;cj袽eӹR~}Ex*4ya} UOLP6}CzϣG-T U t@%ZagXt )H!Vȁ=JxN˃YsCoϥJ9 b%SOWtdTg! sZ&;y^wERLcȘT(qj!n X͖g56L$)$V)/J0JZQhQxՖռw!!Mnցw^;dْ+eŹ_ג-7Y 癠 FAs3Sj-c lS\m ЕBvd(Ybx_ΞxSqpqO=4 6ZSX|uQbt) (i0[ɥf,d {JT!KFUDT, e1 LN(-+NA 6q4z<0Dx=x+tDy-SQ$=o#Kg«N9_Pu7|#ZzK}Wo9oOke 7 tԁ vvw:pB#-}ftIZO=W_^G$F;̄H%E Z3breh6EiUHMi2RRpܱd1Ԩ.hYF"OgQi,e%fݹ&/S:+S ƭΌni)ń* ~ e d}d#]Kڅ vJ=j (rT6Y mDC?R f(5xzɏaE8Pz%rbH$,ޕKPc 丗Sru^nDXOVDޗD4StږTrD\^H.h}wMnEfڭͰ:6H%_L'4L\gih0m䙞7t4}!D`LG_ڒFLT}֞ p6Q+FoylmJ{HLJcbl֝/HVx(xFt55bxI2J Vwgb2S*~q^It -W&25ԃW<5y#hJ)u \=J!z ȄlYiࠑw7PHDԖf0Y|^PV%DxY[FU8ke]^~]s7e^)&<_hɇ*%=|xi2IPGIj-x*tPi9r v(Xyz<$dvV$DO"K#^%4G<ڈBXBt;AUAU!S~C?dAP L`o@SdZrXBPϫ*XËpu)wʡ큍<7:}Yo+GCOd=;ϟ' '{SnN OQi }3|qJev=-Gsj:_]qQ#s]esyRK[r 7}ZsS>ki+<\זW݇vޯ\-;\~0|JϱZ3Z=_^F;/&ܭ'_(8g~/XxꌺىD!qLf08]|7?\'3*BN;/ώG~zwKsK@ 檟?[]fGbqkAsA#{ :"} [t܉:/ޯ {;~X:N߯s/U5dgwb0 iEǜNj_kpȃwe f8:i'yt&_<frv}5Hswم_p8l@§'h=EK[#XFK7Wa~GҌeOE^lifsUh~_o>sɷK%.1vOhOI\?mv.X7R(rVfDetz }K2u<@ /@Z K`^xڜz;rx'!/bdiT߲͆\__տ~]O#u 8:/ofדxћ_~j֏'jӨBܻYluNFGדe ry2j́8cakB_"'[Yg} Tw.S,/rr&iB[K>kۆhO[Xô!Tel*dI *Ȣ r[w {w ٭eܳ5gi@ֵ##/4wb[h+f ]1'oކO\6vϕ/4`fn{51_ ISA_594Ai>$]Rd%GSIYs*Aܖ.YI/#ʆoZ^Cu_2H 1KCtUq4#e0npVVEJa-Z_m~'`sd/^_=I2xvPR7mOޞL=4?ǻgS;<̪O(/w6 ed}O龓?K& νEжș6Eetp֞1 N+hP49{eE=>I|CO^(0fuusbKG[[4ɸ@<@(![z0(N `)i@KVGr 羔ʻtC<\*=Vةכ{v ǓxV9mھ?ii3ShWFiy`E4a$WX⠢tϦ?s8y4v7xh ָ.Z} K?^+ڈ nLh kd,V6t+UdeaPؑ`r&\ł+R+?L*UZ?D\9fv1%L3CrY'0x#TUr>D\gŴBhpEr]4AT+:H0 W/Wnͪw' o]ulm]uIn?ҙr Wϭz H0hprn UJJp0QYW$x}: Per"CsL d]`Ƃ+R++RWJ(]DBJhpErWZqE*N:@\ifn$8%Yݡ1)"4  Q.XֈcT:0}[{m'6eEf{Yx6)vG4Źx9P8 )P(32Jڨ\%ʅoС0oƴHA Z-c*LY r)Q)Udby!7Ng$U^ݑd(!ϳJ"SXިT[[|]֬̍PLJgʃ<*P>b$s"1 A\+K{'ZMHcXxfQ.h6I7L u+'-ԫpS\9k WVTy]TN$A4Bz6\7T& ֬z8aJ{U}U{)&ʡC\As+kW$ƃ+ "\ڭڛTJpupEW$Pv"R'\ VҙphpErWֺTBWJ[]D"FF+k!\Z|RTꄫCĕ!'h#fO;Dk!KPl f!1g5낹VQ~0!YDF .LZAjJɒqF1Hwa$WVTjpuFo}W^qE5W$טXpEjQciqS.&\`-E4"ł+Rko]JW`c1`Hf0Hja dbUb'JmJ7\_\uTk*aX Wϩzn,"\`L4"{ZCt<q%,0e#  P."\ U Wq+cH0hprsTlJ9eD+l]4" +R)DJs![sËEӿ'Vs˭SrU0Jrk-ag7VPt,A`~ ^a@,&RɆ1bIѭfEkӾTL,1|`n\B\}lF*[pfX;qwdHf:+RT!ʂ"•ńW$WXpEj6II!pL u$Ɱ8Ƣ Za+R)]Ջ_ c|9=N%۷T7r;wT4ҰawOznspE-W$XpEj+Tip\ f+ H6Z1t\ʁeK $s mD"6\\cdqE*uuRvJqK7ł+Rk+R!rIܘوS p#gfeLLrkmXDF&"L\%b4_b 4azG`[K'X)aM:'TkkyNۚ_7:7R$"MH/rPʜ!TyYpD%<\I䢇XcףQ.KQc=$AT]>lfg?W?/Uug1WjntF˧e릳^ ٧8pӘ`v4֗F2Y'+fC Yr +vdCWs?(Qh:zu5Pztd̝'٣h f"sÉF0bTzv3XO-@+q|ŗ&tÖ́;|)W^[JGjl>|0U<>U -;oebg5]3S>0>a?лytӇv%MXG?^MI-iz(͡7rTu&uFF8* \d, 3]OHYe+$WM|?k4\G69"uWmg֫]}@.3%G,U3\O:GSQ5fOBJ2NiSUjt78^US~(&dR_짅yۍ+Zߺ1:R#XE;N:ugbjAEI6R:X7'ZZj  9PlbLFͨw2R9E"gvhѵErټ{w}I5K3q"RVݹ2PRܼȦTse TǘƮ#D34f"{v[T rz7^D4!-޵DzY+EGh7~< @ IWDISCQ_ xB. Y&ރ%vwh0ZGx( b uᨢ}"iKmۋT6YR!]ؠ8 ,sbrf|}>5$W-wtM% <RIIT1ڪTw$# <)9DMѵHI)%9įZk|0 4v<ȉ<'Xr!xQkUЗ1\m'r5j'b>#3k^,>*CUW)6QO )ܔ 3HT'*Ht=KF Ȏ ў6d#oG0LUȗBiE EAac^5TTlPtԠ-;y9xC9ӦG`m2~0P6*C!VeWGWd$ty6Tp Օ);:+ Ņ ƺz{EK5i0o w$XIiޘuȆ[aUh]%vdZɷV6K)O-` \:WLCjD*Y0VjLVdVfC2!I6p.HKqN +Q ꛢDd(MJt*TS!z@g d,̼޴X n ZfH57]\?؂1u2r09 ȄA[rvg- eJukڃ A'b΂Gܬ ݄Blnݢ06K#B`@ :P%@ irPfdb'!\(VZ hm e*3RPHq Ls$eIj ڳVywD"=dH_(hT qj3 R +WaD5=() E>R]Q`rT^7 & !5PA e`d."2Dbn iXGjE}B\FФi!:380(cr3RUNJY@1ͫ2 !pBc;_Йxa]ήr=5txZ q`c{6.0Amz`-$ >:%@uPiPrJ7#XLD=$]I QUF2bx蘊'8$;Geg:#.(Z R|$ Ljyu,C Y0#ic偱P&>yt_ 5P2&@ݚ2:@vdm^,:?jPZ5%w VGdm9;77lTbrS,7XktdDHc ѣ.ɧy d<ŬuԆHTʗ]QK|̡clGM玺a4(ڽ,Ck5Kp-ѧBhgSRC;V@@ z1^!Bkڝkh gՌ#{t-- Ac8JA"cQ:fWci63+I@dI5J+1A2~ȃR!*8vGyPYǺkU=`Qy0 !SWI|j8VublWNX- sM{X!gѝmh XW4#[k':iH& qH._u܌bp)+ZƈilnQGb$۽RuN` tB)Ѧi.bjdМs;c fnVvZ+k֬U3| ۸f<%!&ci@P-ŸC69yYgy~5k!|ƅ7 [c땀;*C6(=< VAP8U9P(-84EO(IBjw4CS3&k|TH֞f*JOQ!,6\b$j}E4Xa?`{yv_x}.m E*7۟n )]a/~w[j(xrc\m N w _8So7/0z_~~|8Yjoz^\~MڿRC'0as{RW㛶gWw}<yd3xr]E!涔v>oW_i:>:zhoQK %gd`^?tuz1+ W@^j\Z WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\4\iYX[jՋy@\;~JՋ4\A[++1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp {'\ cc2i W3bz+ bÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J W]vI++1\1\;~Pz1\Hh9D1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%Y W;=(̹Wp>vV\ Ʒhwi^ 26c^?̏[? h?zP^a/ЛqDW0.vpRjujBW/,*!ǝ+j1jRGHWL(: /7Fv?nፅP^Bٓ:}B܇+'lNYNvhEoapn?CE+ݝ5~[oElP U:yn]l9&~> ]{۶P\`@i![lФX L]3Dɉ$K6e5Sԉ73|sΙw߬BW#2r7@y6s@c` y>.f2+@2]xZ l6x򚸇@nR vŠzթmCo{p2!/^ C(}ӡN@tU#Af1Qri QIزt灪WlZ!uDfhic6aL>gt={6XTYEHϴ,sJÃNm1(©noVplh mZ7Hu3̈́B\BW -MRt9>GJ󵑪t Jpi ]ZD*$gHWi)jӺZtLv۞cP|t5PP}tXJp m ]%T5J՟]ϐX>~`N: .?3xS< j3+C>mzPc*>&UB|J(;ztESMMgU+ZCW j<]Q9C6YW3Jpj ]%Ǟ|<]uCW_&ZDWXXW .k ]%: 0Zwt, -|<$úg@z5_[q܍bxd[Ƣc\gR?|U1h(j3 eY̘C:,t#VpjBZde$OHpm!0K|p+#ĸ2!`(U 9Zu݇IJgHW +ҀcLe{&.íYRtJ(\'+MP6X8 hM9Ս!KVb՟Ğ]/P:^}\usէ]fѕ8DGWzLyҭg"Jh o:]%Tvt m]%B\ӦUBIDGWϐ(Ŕɺ U{stP 3+F)׸EtEUBU*gHW< R E T(ݦ)y{W45!ĝU,ikxs׉(*UE;\vNLsQA I;Hb. kͥazB[w˔з?hR$'>y_K謀jQlX013w,7}m3Az_T$((1eLkYA.8dSߦW9Uٿˇ|=ې,y~=}|u]kU_~^]27Ǿ7I^.^吳 nDz@,: *30zкyv!soʌº T^}_el< *Z!QSĸB;΄0b'qQ$qLBQQ"Y^T~mnSgSL2nc7n@)҃p1@:Z3' y$4ViǬQFYJy>0Kȳ`j{b8I(Rs͚leR3ţb%0ĖoR~hu7m*:a _=X|bv/|cϩuJ^EϗCnB2t .sjy%axAEӫ*w)$hE!Z!* !m *xJ+ˑAZ[CNX B8FBX1J9Q68*HYP,+ zkAaeG`$jeњH"H0;e9L߄m:md6F<ͳ _`MΰPiaU>TNv|{Va z5q`knx;N֔80Ack3j1>,{1_ʝ-[in0φ۳,Ь*rYf1.fIN"Z2ڇ7 ׳0*RFQ\x`_ԃn/$fmnh5,kze\rtcDhaqHer@SRaQ{`(cs/F Sg{vƗ }7ķi1Wzu^2 qQogeQRGmcwuU[c(x!fH x6HF_jR0;[,1V٭v6 gqTnHꗝ"0ȗoCo4Nc}yoPFJ׻54rɂCͯe>o8̻b[o_zt[/Ywmu{ /)n̴~_nfEʿrş ƹe f2 tz~?W60 &:~X[Q̫ J̶ym16 },YZ:]XXH3+Wjdk0q331IJ G#!:szH's﨓Btay+2b#Y-)1Xi#FQ fi˥DmJHe40¤d&#QHYu2LwZ D佖豉hjҒcݵ3=+=<Ϧpaf.L,cbks~ T+nY{ce~\ dz#μ|z> љ'21 gkgebo5ZXKMF&d:-ce P!T@a3lhl)Xҝzf7 E# PGU- h#5KY.05!pNb5H$Z Pˆ*LjL2B`"RSFD,OL2h#2&"]Z='ָl2mwyGL_ V9NԛKL=!bCLIOG45g ҁ$b^KBXF( .bUc"{&CDiI EDpC@G,s$ۋ J\Y],la 9iǚMz螉fG~`S.VNV2'])k3sxXS4r|rUe ;HR ';I6]Nhu~]%iJ-k֜آu^S^lr쵲2a ^plGUGX5ejgC73IŹt|gjFkA۹0'rab|o! .nK -kK %n{ N9/4i5K8PaPƂe˴WZ;%bܠysܝy2  G.3w=DЉIĿ;igRUB2 'Zw^D ye 5,0F{L"իH9KJ0h7H%`# !@pЈ | -|shw;"Wָ'Y-&ƅz޻*,@0Fܯ*[C ņx²AS1]CO^"R T 1. p"7?,,n$(&+,WmqqVeMLI"e۾[ȯd?ei"$%gnTP2.Dj50|釕F|߻SEGˆ.VeHl-ݥ.+muTU]4: >?nEc/Fx>I.י2#S2s_>~H?߼Wo70Q^o~z#_f:Bˣ=WEM*[MۧhQg O*o)+ah~^Yn]5 Dן^a1[cy}9AW&xdp%}/?ؠWnE/Q*Tʗ4QT<>(y[)-*~<@;N5q#q}$' E F$܁SCڅ&X:ZcFQXB!d,(j#=uaIyY=41ogjȍ_ac*n]Jr%r RLZu_c8(/2J4rYg@LxtکMtN2y2hAy)eGtbYxkɼS4M9qhnvWsu%p '0rot=Gϓxh,ӶM73aSr*iS)dMw8B gȐ`$7ϲL K8¥xR ! t_%ld|XA/;{Y͗ Ƨ3~W^4yIm¬DCj,\*rIi9NdBUCoAwgxMb1Uhc~kk(Z |7XT|oblVfϴCia8RhuD7Zـg > cA# ո7=WsBl^̴C < LQY.gL Y&F)KuK} B1nr-mfJ3Ug6Jx?9& ÁՋ`}onb3˜\̫o~]et:;h^RS5!Gu;{r4h4àiq/k̶SkћF@`W)i5J&)iԥiv 9*%lC r&*ޅT ȬSe sڟ]> ;2^FcZAK7RNr2/>$nEgz2*F^0WUG ǫW90 r񱠺>ZsB94ԀIEOGW1dy"Fh,P<`)HAV+TYu,3m8ߪY&-ΐ,ht%8i\1rcJ2ő30 ZH0E+jlǝH/f+!M#O6b<O@&ZȍYj߹ AIS^z&O%`&]jHKlDTɄ2A:㜑/Z(|ٽ3yV ԛY3؀f%=&`)!86!R$͵&{HCR )=)zu"/}5JX. e.HF 1y#u2 rh gYױұIgt1iںLHZo>vpB"8!>zNNaҠ40Wtѐ{BE Hg4h*9^U `gإRO&}JDn)J<sdu%=2$>ap.jfXB $|K˧̥Y|I{i8 q|>AiPB/i4Q>~!Sz0mwohe&Ⱦd;Ɇ5eX\PKl--R 6.=z$H 1Ó))S)j82EJ+2/Q}8~Ѝ56_\>FMH\6,4qHk*hu)Yv83g}sa1>+ZQa"E)ܑIo[AnEܤS* lܥp"JD"\}IOBjnUy5qa>HnӈӉ* 96H":&LA֖Jh-Dž^D TUdD_4 ;L*[I/b9& Y/cʊZM튚>N4DӤ '#߾ <8Njd'NBB%&'%ف6h6hּZ$zmN99k.s[z*xkGn*9U+lyP0hZP 1ոTD’הּr&ǘC`VyZl̒ 2= KV}6g&E͵msjX-63vPBo j /%*6;/x edK=ez1p~ ? O>;-`|VEibZTtDK%'2V.g$`EZ)Zlie( AHQ̦5 $&oǬ\3t*nx\CfǮV[Wں=ݸ 5I] VYL1s GZ 5g#Ḫ0+m<x rG e2dkbbE"7RpLNj*2$:+jlS ,P,bE4--bo7xhw(f,yFDy.y9(26YrY1{ Et3F#Iͅh@bg_D I'͍L4OpYVjs&D:]6r^gY]bo7F4-)"F xV"R"0Htǐzx0sٱ=\a]Fa躘iFn |KlNQw냵)XېVգǟGR0lC?qV!d1zf Nb ЉȲMz|A;ɷ_Wq츅\7An_ܱOi yjA/5?B&u)gga- n0)0,)DQrQT%!& ; Bg.DV&n#Ȭ^C5q9 E( m/cx6N-XLqē^Z|y_#|=UӇGPRLG4Hk˽Cy^h9ل wFjid&Ko*Y2i0Y#kx]*8`LxPVB[HK"CK  O3̎coz֚<7c1.J)E" 6ފB/QO&JA}:u2? czl$-Go Sh2TK&:Ch4QsUl X4ﵾ )}ס+ &wڞI. 5,BVK5xQ.D3-f܌ 4%E`ABo>9i#B #cʥ4N$Sl]BӪ=wM ۘ?oh9LEkE= WTz0@2(*R-o(})' 3npr2-W2"RLi?%rO4oӍʶ磢m$SMi \QDڼYbiGZQ<~tˈ> 斎'x3N4x'4WK /Q'iȩl|vF#n W7wBꚕH*?"yۻ6iדDuͷŬuD]lޯS9/gp28pOv-4ޟώc,_|9&uOjaH/$x= j[~} =j:+*m1윽ܼBű*<1owHƓYW 1ZYoJ{j˻+)0:[I2Qz=g_ua6?=&J8 @,A|s! ?5iNݕB/{-ty;ú?%k&~zGs>y\\y`x<sZ֦pa`,B~.nsGZo[fwgϝ5Xsw ,R RX3@]{KZeg&My 8O;=,_l.vsV˒_{Ӎb$ty7_mujNL p _*Ѫ C7QggIu6ӧ5_PWg^s،3Njr$8#cNs@witJ@PF+3ffg]mnL!..pZ ̺EB]imjmki\sأAGhi:ri hk>OcrRI sMP5YBV&IGIϣډt˪WP`'MلRahm&O21N Fh1tc_`θgPj}=[&=AA#}1C) 2» CGzv4{vxtVc0E'<<"Ǯ,{$d*x0)*[ڠgLUJJ4X 4\=hAiAVjc3l&bq]-tddw~ggqk&!0k@:J has Sv. :trk ػI?+e'~Jj.'W!iFzJ$UڤfҥgBzMKYeUWсveDv1$F:JC}m绦t|9,tj[6,/om7ݙgJ{8_iLAތy0ňjSEϭn.-DvHTZsoߪ~UfulVdہdrDw2,fh/k_,8`4fp}8^fSٗ^3At9㍡UB+ձUBIKWv=HI;<]mcq`.=0]mVCqѕ܂dKWv=L Jk ]%1th5NW %-] ].h]卡ӦUBst(1F-] ]Q9" XU+CW >zN`O!V+9i ]%-A*$-]"]q,^t}%'OxM'|=$nIe EbpA4/o?ˆM3كΙsLr0;Xu>ǘS1ŋw7עS|8OK`0@ur?~sM{bJ/@{zԞ1EjX쉜n. \,}^__ӕ 6͹Z|\ E7Z;cs!S RX"2|*#Zq:*lI 5CW &LEW 8NByl-]= ]IIkR9QWƼOh 9vJ()kJI&K`ec*+@KѫJKTX7\֘ -?7d eu'NWjîWHaDj< ڷC؞* %FGjW[Еjj׮NJ!DWbJpn ]%;]%\tutE\7:3CoZJZ:AJF֘6']J]%L;]%S+`Atn5ZJEZ:AXT.b3;Tu8;gc'([ԛk^Qw`4 2Յsc̱|<ų}%{My2.cacpF{ɱ6$Qs-9˝R 3| #ثI&ygȜ(tт&`,Ս|Zٱf %ovD7k]%9/\ݘPӠJZ:AR"DW VM+@KѿK(nJ#)h^`)Pc*=&ۡ Pcۓ=Joad`u[vh遲6Cɏt[ڵ1U7C/ 6H%!-] ]Ӿtn ]%Z·JZ:EJT;5 4V0ʯ@ߠ7fB)My,C#傎ޞe~8i.6ml]fH?nTh_z 'aj"9HA?Lz}OeT^ك]-lj #|I< ͇S3A*F_8*FKMC%uhQ&V\," ] ë8 D%ErrAӌ_ * L :OT; f=^QaM=SҦv>4͞rx>{ jXppĺTcj?X+NG1GSV+4k ]ZpY"]q,i} Icv *WܢބqѤ׫XPNpycTeB_<ܟJ[>EIԘom|AwMW!`!6%¢1-%1L 8dRN)ζbU*1t8ztP2 ҕ&201tJh9:vJ(jJSZ+LYsWUB>UP2ԾUzյIstR}mɡ.Ϯp%,g7fx/~tq?uR=;6zFy.9C6XbV;DQE2mk07B;0jx3 f-MIg-cZُflm1hVAwP"PeW0G]wQv1;j-.8-Qp(mn]A^t}Yt% Qڧ•Ӷ[/V!hozX:b&)L#dtݗggY!~B^syoUi`j3xza>ʧ\ u~Q .:[ZMa@{y Y  )Oc.*51Vy} 0/ÙHJ>c3_T^oi羹 ťYŔb8j)Og.C?u)kOTZBݽ`AI]_/~P俾 îr\t)qzlT5L)Z}Gw{*VXUpbvĩXUW+N1LN랆7]ssP1C%U 7Aͦ9Za 5,j ˤdDh2/7Xv0"[y׮SZ9DNlk9o!tJ BwWMjwW]0Jed*uvL =E x$ )[X1c2b=6M Y!-+Eܓ_Vȣrm'RqZq%X,5IWC:ߥ_Pwm׷דBmF`&EA. re,2_ǓǎK/6EeWM "n:Ͱ rx{=rΠ9 3pFV4^ F3]^8nGpe.E/`â+I6h0Nk\ER?(.s1yKOr:CsG_)t~ꢊ]5̧ub5鋨 J\~< 5ή.cfG-6߾><[|SSOgZW.]_bBX^,\߽f7R;MiE$Cs,9&ʫIJ3 ]Nҡ]< W2P8eƊh̙:ܭA)bG.-(gJF!l#Konll# #vR v8/]>(N`BO;˜.3[y29ӝWܹߛ8G|p%q}Ոvm`))Fp" t4 Q,&hՈd =!"lfh CR(Ŝ(a"ku#$myOUuT4e쾿XiǕ"G-hS.F6ݓNowzg)w=Տכ,k%{{}MTSLU5YkJ='Սoí{4V?Yǖ|XZtxMb_o>1VR& -Sa𨢚+&9NxQqI>|=3S#7Zօ9 CRv^:fܱisϵ2]phM }ա_IQu>9WRWRَ9W8iXoI@Mp^nxh08ޘe{oHaBR*%5}6\s@|+%*9++aÙc4eP=LfiJ3Ӓ G,Kn@%Q [ZcI %V8dY`w#>BQ0Wz#̚'o_ąLz0O$zF`ʠH^#7u'nsߕ VdkV0g.alXkoDO |j6l:SDC'&Q2jiTQdEOl䰯eDLh1=zGD #\T,x* vFZ70q֜,$duGVQPGC U)nȦz̝vX-@^%{gqoR0#4'(xAcfQ09[5O?w2;t\-|{O#M*U;1嶑ײOג%t-WE^N'WH?LH#0ZbJ`f dVSRaPrFJV&G #{[+ 9b;N\X ђjɣFV Rv_!ZaQɞ/q} ㄙ/5>=\Z:]sp/=W]4XM[1oEdy.pTL!J MfV /u0OAym̎ Ŋ )TD2'T4=KI%pD0@z'Ll6qw&n$0 ́lv;na{昃uSc(K|#;=D/'Y} TB2 'Zw^D ye 5,0F{L=#e]caz1 QB10XqH ηbud}L|ZĸQ0+,ñ3d;COpp)Ş p0Go?}(Û֎SA^AT20P?o x70u*|gE dvLbLUs;,46;fsm:.AnQW_m5 -,#Mm hRI$?4~lazҳ&+Զ.ɴ&_;/fg7˂-hĜ/QNϚz1KƑb4|1齁7abI#1~H竆!r0ˋ0qOdz[|<^v]-fkN rT֏=QW=+-JVHxRU(ƱHqp1nf T[`.`rOG'om듿ۻo?`N|ûװ f`l +I&{=1_|ianh6CON6㪜b&qkev[n@?_|y=? &I/>H-]9 w_AU\JRX{q:8Z>^"UރN9kĝt$_߄aoQ- w*U -;֘QlְD Hml hpDHPBך"q G-7(2zYq+(dN3*.&(ۚطD;4dH}xsXջ< >XL-*+5rk+q.`9VΦaΣY[FH(򥖜N)bxЄ'x:Jw.sMڥfNI52AQ\6LrsW߼>ƟG+ėDm6F5b_*"0#B8/%QyQ+UnH4"A<@ɶ>e4ͰtA$u/ª 2"i+Ei%aTR%5ke#LJQN&JH阉$/fXPMVԁHH ÊĬ8 AHRVa S띱6L"^ˈiDk45@N-k!{ʲ) V KDD\rڔXE'5l?K BޚN5p@.(|8}]`@rPNZ)r595N iiA]KpL.ŔmdJ4ϥΦ[ChZe{oCOpy\uK{EbЛKrEvaYI-ZF`s qrw)KVTm4(>P2zER:飶ěUԀ"TndFfd'dl+cW,dXX8(="I+RK77Λ͆of8?-##+$DO ))]b1:@/ǤӞFy*LCR'ZK(IItYe) Ӂi]I;`.r6#/mu"3j{ v. Ǹ3$U\i@L0EAcJ״d΃#88N#Hh5> 'H5RRx #JudllZ=$`ܛz}Ǯ(3#GĵQlTaRXjom"\@eup7V!PA7Qb%#1 N paK̘;`.rtU[ŪlklVɮ2qŵ7Ax/"%yFyPChιVǎxaR6+R^Vp^[c}JǑ4nNpy?R.q1r|򆎌!R eCfu^scctR`ʍR2,r{{Xhf֫ V!8$ahQwrkv@hDfrxb.FYȳo* i`AHjz ?'_H 2La|״ljB][}ʇh.+/pn**QNQu'נTn7;*>x?.9!hZ,hh6$<^ NX\W*?QӦ[^k֞L)_(%0o:r/RCxmΌqkkTJCnOxtXk7,!d:xx]Kx:K#m hiůS{_ Eq"ʢs3Gsa3KwJWKs\NoD]9Vj*_Swn}u*?oJ?~'G@O_j.x4yܗUbT =0 즻V]jwh66y=i& n׹ۆLR E)!,b1ocf&0CsfNokjYR9fWM;s Bimf\H$kӉ'; .KE+L0睈Z*Jy՞y ok(Y?wF^" PR 󩤶Ř9(نIn^Z.{U.53ٓ,y½Ǫg6}MQJ7I}EK3GK4'MdaHKŽ7ևR>lim`{4yT."L.{؃bG.8~ vR-EEM@(K'nK- 9@W_GMm8-\iB@ALp6m^^ݶ<7<-UcCw7RWQ4w݋11|gv1U ś_+m#I̔g,3  tFF&)݋Y@yAoHmچk%Mb BHJkd: $2&A,"%Z kшQK$ Y1ܮO7 dΧ&{c:l-:[^rͿi'ks?o)+i丗ۈmR8=sMWFXph%J@f9^-o'f",'ҀVvwPJ)u I4#n,_|]z|oiQn8$K]bpvkz6'-77Ly'M DϔـH r F@%&^F4v\O\] *]7VuNzzOo{CK{+0mKuZI8׬,3);V&xc47 Lڪv%D' uU'ǜù Z8Xۣ[#gY߀qnϏ| sqvEgYe)#enR~P)2ll KQZJ++)|BWy-Drk_=';f+42,qQ &QN07c؋s 0#K \୍a1,/`b@Rث5rvhɿN(CSK-'؞'s_6Shu>BߗpͱCGo7̜Qh!ذ)- D wX Z̜W>T]*r8uUՕT;TW g$\t@ ?Aq~MHr &{L])qDό~>Hosz*0=(Ү$/~zVȔ>#5MSg .GJ#;5մlAƷ^fF෹#~I~̓~8M&zשhY3w4 \lɻ Qޣb(H]BQW\#E]uJ.:ujuD`p6ꪐ\URwyQ]o9AY p6`V|BoG]KL(~p]u,ӎ0G75ߕz>{jt||{\Q Ul%ϕ^VKolR: 2 wvsrEo "X&mEHRe@R) kD"_+8+2`%X7Udi1[7xvX~xx'7 t?8y?Xꤨ .6q`>ڜ]/>hM]V!3R~YbUd7HsC6F%5!&@u(`I]]ܟnBc0xw?({@}Iޘ+[gbE|5Z ҵ[~O]߀8&S@ֳRAķ ~ &eFf!7BZ\bJ:_)6?NR$5s\ړ EKЇvhufRr)i926%ނ xf1ZA.{g#m[#gG`i|f~Fpw(<[YM6 6" Hz v`+i %_~ [/z?>(4-v} )|'ܸ$ĐB-nkH2q;&⿧@)By5R2yQ.|xHe˼I>[iK_;L=~&va{)0M~Z ؞!xqpdЂUƮ=~WDn(^: |~/qz(ze{Ĭ%^*k[/+ۙȍ$u; gn`~Cϼ ϓB-'nv乭 VzCT|Κǟkt9-J!ZE?Li{#OTVi}ҜeAJ=mxې `Y%!JW,KTw޶v޶(:m/Zd!X DŽuǨ$SeGF:+ZJ+.rAX"JyY2k)U,iD {omm-kj% }B˖ /K$-jrU3v%$5 ALЪԳy$ kKUkIiG`3EQy%ye#Zt9&CWzUpBOu=:R=Z@PBX^[FeN}r@{n1293HTSeRC R4)9v\D7׌wXWG.?]ƽO,^'_5lt%c׾'̾${]Ï׏w79VN,% }x?x}$UQJ32yH@q0N"xdF "\(=/1)toٿYr*ƒw*\ /nٿX8(o,ݱ }SwKJ>^~ޯBWWŏ*`knewmLXx?Trunr]IKŘ"i[I8 4Nɮ&Q*DG0Of'+դoKԖ.ii ~~loXO /׉a4+a̧$GNzշk#d:~1Oq&:X~6y" F$܁QCʪ(X:ZcFQXB!d,(`#=bCye=ÁB:$/q#V8jD)䕉sB4[P$F)%#tTa8?hTġE#¾;;_FܫC' jpy)$f }>Y3}.Y3k94kFR!kW5+,8|搣WP&DhTH-SZJL)0">p.^4t';?Ys@" LK+)J`kP4 lp!j'BY”80t L LMGoEỤKprzcU᳇>!Ri7)o@p*(2\9S0nHj%BvP,_k^/?<׋ܰ?Ab7üb/EoY!@7o9J3~ '4U%70(EKf-hYfoF# Vda1eÁX+>Mtq`P\Ng9~!H[`djK G'A3{G ?'<vkP,Eș7&2QCLPp%,zb#ye;Hx"~p(3D #\T)0i5r8d9T 9.$duTPGC )npSl=N;W ha[0 ^И ff*!'1֞犏>ų5:v!yk(G#}2Y=HW9,=g1aOf\]n\5ӯX"BR6`Z2XiF%URȽMぽ~}a61b;=@lWI=Qs/`15.Д + U;0H>4=}@Ķ80|H]NChcxIP\堵i9E'5L)  )Y*Ȭ@I O> `2&ȣJAH̊M5ի@Kj.C81mub&ybA~ɗI}>-졸I-ZF`Cf˥Y"0hQ*o^>d>Z70 ()&=E0飶ěU  s32*ٰ0 Yf,d+^ʐ6_$f1SuQ@ic XiP4$.@=4=+S@]b1:@/ǤӞFy*tC2'ZK(IItY@e) Ӂi]I{&87#Ny(mt"3ju. ˸3$U\i L0EAcJkZ2eȅ4*?  Xr0#ǐ؁=r`TiLf<&xXݙǾ(3#q@ĭٿQlTaRXjom"\dup7V!06Qw0pEXA`Ir40I%eTKxpZgQ/.̸\pqmEA fmP#FKf%H Ĺb!fQG16 xx0sѱ'{|ö[=@2@MCE5gSRosޏr-TwJ~G{7ϑ[\ (24P+CҕVSWfGQ:8=/D)K2QGb1L /^Ỻ@q}Q{K]!3T6TʜhS> ư`.pCJ+0F,BF<)VkʈhA #(H8HgWϦ.I+v*9ޕ~l(e9x6r2ڛ}ce=:g(pwR1`1bF;Xk0`k8U hD^Xn 0<0 rp%13֧`pb',ȳ9` r&MyB^ל7z; f8e&NwK m(6p`Dplx-1F9tj9A&DY"#Tr)GZi%4ka;}f1_>CMD'7A(PEP-^:$|L7?{3>1dZdh&:8o|@/tLn2mOl,$լ=@0*KtZw%]KTfOYfOg-㹖亶1@;Sz̖ N%c-Y൶%4,槩]ry7E(:L־@v<ۗu?_0EIgg!ӔRq>uqwoOAΧyz&>`PǛT=-}f%_'ࢸvDHeU9}n6#1o5BV׬;Gn\qi0Rm))tеuoǓ>ȿ^?N9o<:%L1ܷos"\^K#weC9mlH޼3>6}PJǧ' tBiv$uIs0HKO~ n:7BҮ26k'UISkS#}8.$0a jMɾ;+{}ݷw;z7)Z~0%dz t)Y=e_gy''Op>H Qxm05½M|rl3yi:y-O{diƆno\ͥkQ4w]݋61|gM|ߵGYTD2rY.fmhߟX =Õ=zٌ黆{ -кdRIJZ{8jlp1U  Asy(?2c@ {YT0BrL0JX>euu#Vº :q`U)a S*%)d4&j9Kߨ@%zUa6:$ .eq")$HH*qs#Ņ4 :v@:6 .#1t+Ni4B{*Vڴز4^<#Et߽?2&#Wu; }*, Z8fX^ymmI %ȗZrV:AfvòΫcP[bgKڌНU'tc<}1͒[ Ja^i& ^ w ^sXe{ mD^>H&"շiXX;EgC1&0)P.͏pgdzmWW߼MO'="IMgjw8P0/ӁŶ8xkFΚC6L5i*(#Q*F:kꅒ+˜R; 9MV6J!^ڠLiL )& Zik/W+mH̔W0vX`;x #nMK,/7xX<"K[b2#8YEM4)BȌ) T&-j1qy#u6HZDM:FqN gOyT+oX$xr 1)NivXjyVay%xHn!5n=;:;%h숶&Q-\*< VGr"2x8Q%V^yP{&ٞ%S!u &6P!:eL!O1ʘHsD#!Zk@@/T{f*ּ'@d]_'^xt\`Ƕ{w"Im֩ޭ HR% aO>.ݎU(~oiQn$r.vcB99^Y:K6bO,Gp241(=S:fUd`R1*`d>P3IP[6 ʊfUڞ-#6i" 77zBnBt۔CjgK22)F8qD Qi1ZdV dJШͧH×&sK MOgqcPA#ʁsڂYWW Cn 0|6l\cW,K Lcjn`.Ss+57 ZeϽFAix_s ܐH:J=QW4]QW/ J^]Au8rxO(elLr5J|&k&9!SW  ~Zei}NkTÈ8$W^HǭP.B}z>as!ay_ɞ@ʸIסaj3Pel`!0W*>mѭ~~U[!]20)ΰר B+",(y2"ШyJp]QW箮 Xގ2h"G \Ϯ ʞ]Iui;0TgU QWsWWB'^]u]]-^\0&?}S! ӪЂ8 %;+ѫ}3/s0; \ͻ Z<{uE(^]Au%8%vESQW K Z 箮 J%{upaK[: \ Z箮Jxb{ތ\VRWXtwEh;{c-+wqqW;~ ږ*p튺*s&bJB54nع{#Mi %4')Qw ȃtCyVx;'# [؄Rh\mȬ_IB3DcR֦fFٓ ZS?CV,+J/NI`m%E22VNnd?GV@]fMsCtˌaގ>yewbY߽@KL kE6Y !D.1iLhun.;dQ`U:p's Ja{ ZVF. A d򵥮 \. c J{uՕvH]; \kJCvB^]uq{篮0]iK]/{*hOUs77_-.hj7hNvkةoRj練 "oF˴ִmݏ˲5wՇ& R~^byX=WCPWr*3L)f6B 2d#*MLy[ݕu?JŸWmbuqȭp6j0W͝{g?>~mK6n}u`G;#s"fG8?%mSgI$H^j[ 3f_> > k?7._6#t!m?=6(|Ц*D֖i?R: vќ/酞M"biqh54<\EdF)ܚʹd* )x+8ߋSv^梿=wTVO)"r'!SI23c6\eK +RVI Bkc$QC[u 2 !j-}h2eo)tpR$g!q:$l f'a$.Tw^A¬h羼<|`7;fUOh٭b>=|3:/%XH#" -Lr2 +2e%hTU H[P'\#:Hgasr!THr`e n f ^ɳ^ߏFx5=ˊV(.\h󯚜v[s#SdM<#<kx {5i40l\cW,KjjKe-qstjQ 阆$*U+OVR`٢FCKѶ"ORq(9Iq J$ D"_+bTdWuI*ER\[ճZ;,?OCw"m?A77tOl[s'rIfPNrؙdBt%'Us J+7̍ NFضZi[^UB_,ՅTe;"4,WnHJ6ڥmo0h~*.(AU:$ˊ10`=c#D= Kֻ2BMp٤X`Hkѩ 4 gk"Uv_͙\JZeacJK>=r)ň&sȑ#mk Ζr<̈J4{,xC &azĂ d^BD׫|K;^Il6XK>?ᵿݷ_TLB6׎O}mPűq!%]Y6fvzO?ݽw1J(L%Εu9і,,I#3e>@K7EX=l򟃟82uH8og%}ޤMBDs*n'|p?S y!XvC5Ow](݄הd^iz_Q<8{R|NŽ fC6-m('nӈ}zXڇ4ك6HuW_= n L ;~h|>|v7n2|h*ZƮ$y=ʼn9t.u~ot[K!h7>鱭A6՝+f[~$nkrؚ :]]M^{zYIrs?tfEgVպm=gw~ EB+'HF<7+b.mEgYͧ11j^0Xe/Esۏ$ߓu1iteeۮ]Pjlxg2T \ ]P%}jAU3TlA[Λm%3;e-8ֲ :%YĒm( 6BX;c+cOY3-.Oiu.hVrkRZٹx7<]x:X &'3z!~T&eballc$_gc@]Gh{ˉ<.F)O˿TL!c/<$-_PɨVK4d16%(3%-0YƚchND(o_\7p\;c-עЫWë)){w:\KZ;s'yKUM{?vMNQsmxUbƘ`̥1'p\F{彽292OƿY7V7wDAM-m N6567fvw촙Ɠ~BPi2],op⬰b{ˋk5:kIuZHmpirSLʠơ<|<)-''5?p<=1wǿ0o{}Go~xskqw`FXw7`m?e0旯hZw4lZ{KӶ˫^뻼 ~> ZC\ i~t^kY }Q_\+p=΃ n~2ON8EkO7r:\=<-xpd3?h_&Ki:6Hh#dAUh$ A[HŦD(Q'״>N8y=a5 !ha,R':"XR6D4ɒ(*z];NLiv~/:Vjo]|E./&U}Ӥ~WC9\߻@ŀ&/oAB9D~}c]Aӝ2vi^IyF #@=X|<j6qcZqQMv&8bn ԤTÔ1dQ$e wj+!%#D5[FC$D$cKşTɱFI+ 5ZeTx}6y7QhҎy/i7Sw:Vʔ[Z|x$h9g S'> JtN}ZPlIgj!v2 -5ZeJVWz@uB\YL#f2ͱњrj%F |̅#neC*dɬʛ虛ʔSW69!  F V`gAGcTdGEU&8ҕ빑pe\I m$9^̵R#^'0UG$OWAFArRp4EΡ!0yͮ.RBȑPLI^Oeq1:#Y$'h JS׉ GO1D)DQA  KC]Fv!A({a$Lƒ5*3/dǼ3snLh{X_eS Ua9"DLvG@L6«>W;/)>yK_ɓy@X8f(Jt'I,Ne!T(PMBJh^ɳ\5aD4'DcJ^ o2{w%XUN[i:XVW(G#:E:L%D%EΠI1Emɺ(O%GtE )i}l|q{i Z-HZ?O>tXpS©M 'K%/K8iRa0a ~a B3;;aĔH5En3Р* Sq>T.sCLfJ=ݭƴDx0*m8 orsݛaɦG?6O>7@T׉1ݮ@ LU462\u+Aݒ|ebkPGO:%xcx596p(4lH& )]P䐽93NgsVu!#GY12٪Zd`R+^#S21  **wDȚ@H2Ɣ: +R,cI1FZn.̹놭G΋, ɇU|As" m@*;c"M.p\- v@Q T_=$SFOBl8 r 9cG̜>Y8\l^&agd'MBf%KMA6hvv^-2x|ۗr\Luw2]v #_.kt;E;W ޠu<Tbɣqu~+SVw%璒& sR1UQ.T})e+L1HusθJgXY c,  WkWfAj=ez1hhh44>GF쨵dE=1+.M6gR*^C׺[^P))*ljBƄ v‡ [@绖Sʜ;MJ/YmǨm{Ib.r%Ƈl@H8Qz1C*!䘯uX1y#kP2?"k2 8L c_dTc<̜ ";"]="nxeM1ʔBNȺQg&~ԉqq,k:;%}=.nrL%*gKDQ"o12.RD=.>.MUf|t}5>7(t]LgtҎ>,/QwH&DXP2c\Y,>㏿u(]$El\䳂pؤ!H#II/3q'VUc.?슓Ҏ}N$Omuel$ My% )#gCRﭚa?vWEtb4J,͢)jlGy#9ֈ#M %!O` UdLLIO739z6BO<xUNJ*2H3rk9lU Q^w5M95M7g*|TjgRX`sĎrvQ3T5s+;Ou'.\ ?}a҆R-ڽĞj?n% {9^'k=Wl3 b.p6VDjͫ+|݃[q6'O>'{>?7|Vt+>˂ho^I+. Y\qQeU|J͜&?Z>";|L8f{۫6iǓd--gmq؆̵zkk>E^U?.|v\1pRoV?0X4I>%C8~F4Ip(;sP9H{|R3͞\3ՠpۆ}YL4% ;F vAi}9Ogc7^5~{x0Og2_5@ls7md:Xr{nx`cV [_k;6~3{Lҥ%uhV<1]lvuc,҉.<gdWH=r(rSk-g"TcMyڍDr'a\{ŃIxQKEiGuqd9h[ox;ۭv ME ,e0\c氷FOu$N#bݺ݀^Z.^ENsgkc׬H2v=VGR:djN"3Ũ>J`Og"WS rg7X3 =r6{(륜|t}\7O^!;_�<Ϧda4A;<d^Q BBI!D{昫S*  )O\N;L'*a[ie󩿝 %.rڳŒ5Ҟ,,CV9 ~|*'a~˟@h\m6ngU9qc~H~kB䄃(7.,Շ0> !"wzޜWOUtjuSgU:SS$~1ǘ1Tdujzxdti%,6n~ϳ] ڈ%w_zRQq[ߊ\qQK(QɎ̵ʶPY<Q;+H u-jM[iPc6gBy/s`3ڛ핆WG]Quux>q\ux;Ptz8;^9笽NϨ$bP`䑕1JhƑ{usՏd;p j/6$Qs-9˝R[5ˏ?r{nkC(ۼ(m]{ZKIK]9^BlAT cJ#6b]'g)OK[,} *۪ma|*e7;io7a!t> pgn(A$S^uGY0%wӨ%R}Y:Gl\έI>2/S cy ? ;4W jA9&EZ"FKEj#QګTϤRQN0\ڼFF:"i+En9aS%5CV{%z[ֱ踶"|,[;ҲB|19"q \ "*hRT2ERGnV$v$C0!B*Caj}Jά1c2b=6M )%㽧(grt@ sq kY%ql""SZa9pqL"h8,hd@AZX,vGE38$)p"JQ;#g=OA- T”ݱdl՗IVlzW-|HkrV_EGv ISqhvKSmQy K&HI-H`Klcƌc0 V+q d>ZRN)S\-&x#32x5r  52vFf*aag+cW,dc!5fI3nI _,L. 7_5p84 XiP4$]$'z"(\Y%+rL:itJQ-bSE#aHƞQk % 6N0+`!#a:0- 9:FYv< &池vgcWQ`W:ᢀߌ `i=#HRŕ!ɴ Y)+i5-u iU'4C Qp4H$)@q =r`TiLx94ꛪT ;["e="Z8׎b.w:g V{k#NR&s@x)"jl: Ą3`/:łKsyIjd[1"vFzDt`-b=&_ggdW\Tqŕ7A3{qojfP@+bF|coCţZ;bGxX5/YPȵ?"UEA\LяBwI}7=~A;wРE>n4qaO(Ơ ϔ_Hga,"?GaA U,]V `uM!XS5qR!Y+냉kʈhA #(H8Hwl5tFΆASk\>+u֑[[k  ż'דKx.AO(O]^MOŏi~P bX  `!€ KTI* zDZq =0<0 rp%13#!NF+XEos-`˙T6:mC9XxdcOR"k/%% XZYh6t F#|%($<hf= JINxǖ}̛x+8[(g[#'Մ(˸:LYLkb&+'r~緳Țl>ߪzTN~Pp1/˖^8逦Ba`y.!͵/Rlëo/4ρ&,yC&“?Iy/=ߍxrHH fGuelRNĥJ#\[̮te&2s u8JҪ R6\*fQ).lj&V^l0Mx۟#6{N{^K{Z;AmVU!m/QL6XX=Kf=^2{LͣnU[ 5_f]LQw^n=ÛH5)7y ajR]-]ݲچsFGf\X'55 vg{`/;Mt-ߓTY_9ܛ6 `2|nԺIO6Zn~y96 ͘1ELTuiK_ D[("9 />YήguPF IQ) `J${8n0D{Jxy>,.'ܻ/ Ć&jc^WYU~EhdUWnXX4?BXC#w&ѦS#y O}9pc22~FG\ Fy/q؛}yg:B,g˹鷭{ջj82qVEhEYٝGmbk#me$;IJV?ϲӤzvbHͺmwP'Tơ޵h?l`hX+ ,%%>2VUamT Jxu{IHB/UpDBHʤIn-%Z Y Z ,uBr}:&Ye՝Aj^'߸z>'ggr60}@[=B'ÙyRTzCLi Zp0YCoppYP Xj#VG#$K3J1&=pFgk_'}Ιs*5R=1KIhY+r3>PTBeLXߖ<.k7}\tj;)|2*/8089 N˥8z{s1eG *DTh"2+iC)S`S*Ʋv ѯȽh?vR[_gT AqJ { 1.B?x~@+G,ˬy-JiS0C(KEɆdy,'#Ro=30#2ZʿSecg<sQ56I%9rG}8ߕ)jv,$۟X{(]ecw;`Zl6 znl $IIB OqZĜS)Ps4Vq8Ck3 Ȁjr[|xú A>:DsDs#/J"|a,09:__SvF;JC!KF-$3Gӝۉ(vP*@i(zN+ `%nuP\D}$Ƙc29^Q% )QNKQ8~\Kyy"z}߼~/V[8!~yDu7\hNh8rD 5RTI-Q &GJ&r\}&qyEkp;AXP 0WQ}tIt+US"|mNťh&!Ff c.BE"h#ꉘ:ՉW𩨪j)6Hi6?~y'c׍GZIێf;K3G={q=Y8F;X"EbxH&;_Ҽ,)D pRmM!;yM[tNz[ x4qqo@h!\dgHa$ 1 `S Nam JxF,p5>(L0|AEQZ(s>tԬ!#q!gc,1jcԤĒ8!Z" PVz @=Vp.qx ̶TH̵A[n} GßoE@֖$aU[`yFpTOYeP2=YٿOw4ԖR獬+l$;FrϟGc};Qi㜞ި{dC?\pNG|O?~_L|>-8Fwk?TIP{EQ4oh,o^4␢V:ruM)B٢ܒb- o*Sȩ\[-9TDvS!rǍWoj:sRů(Hz_ryl1Do5'::rԿMZKVVBonK?Մ\/o.qc;(&x(@Ho{nR "$.oo;vKg939fz5HQps_3z\ƺ八TLwjM~ơ8SbgƖUa٢>©(]DY_ޕul_! n]? ` A ~ ,5J(6A׾@=kw7-SmnUNժz@Lt$ykn}}׿keG-9r?z)g5[k(~i[&LѾ$*E1k_uo z7ݿ1xYm^ \3so3(2eoFUo8J.2YBd/-)tiGVB2%c4PjMYg(Rޞ7FP{Lu&cH˝t+׾h- V&:?; RZa&[X}҃KԵ]beLhaf`A'-Ռ Ӽ1ZrV>L7pf6dڦbmH4mΤ3dClt)C _;L>R;QE61b5]h0F>=bT: p"GpG+J*&C8=εdCUJ6Q^$TY2L>R'RBu!wI4GUA2^sH#ϩRRnm5ljx$7.(]cGpW#̇5iݜ|OZwݺ{cF6o3dTMdB9CEcOk;k]Vj!G]чΨ|W!'ÆVܜ-2f %xxM]Vzlsנ#Ҋ ;mTф h씱OrNc](- O`fR }qD2.)fu$Iૼx" b:cpmP6ZE ـ!:9ybY YZuߑm pՔ:`AQ@Qg=ak l#m*< qhջހ\U6G6N")4v`k:Ye;SCʝS2gRDf Qd6QNsd|rEt Mj*6F"b2i2y1TJ4bxʨ̀j@o]P?!w c6Yo)"9c5E"Y ݄Q1uK$j0B \t͇ %\A[*MC{Ꙭ@]4YH) J8 ੲ8BpS;KQ@hZGS i0VTL]dcg ")bu;? *EKR|`DV4:   e͜ƲU 1:A db5WVcYuN xfnpX{݋~Jld1fR2C8i "2m3쩶A*8[-O%sɕHWo#fVXه4PQFEUKqVLA#V  X;>."h YyH!~V %dlE'jjUZ eJ?-j~B:kϢ;;cXF#5[A*uޚ{%E/ h߬t6e H-hVsp )qiє}66`_nwh[/O]iȃ~s=3L,2:tt'{gѳev{kgU bj\ ϵ昴5g5rFCoF>5`ʛ㌞ JHzV1* cA5(Qjp 8P`]@HЙ 4!RCb,[ gc_ Y|?Z?uŠxZ96OI"ߑPg}C7 .0iQH0ҺG1IՍ4]GTXd t)clchУ-6P7ƖJܢ4RHǚUHLzT;UT:g},ރ?!]ےY޸' *6LL@&0m@ "Y5ކ tijAyP/FC̆jS"qd=8I= z6PO,yJi]0 u[ ҴxCY}6\E*í&"˱XT@@dJIAy:ubٹ:ՌhI5Sg슅޹}@F(G%oCL,ߟv-nY./vdv +W S-00fz) 垷wӏ?}diAɔ_$[bѮқ7jlV3>pe*kŕ@Y)`%7J 6}Q>w%/RI@/peNmÒ{Nxʋ͡QoIϫ|}EoȾQ_w-/iH}ZhIY;ӣwV@` (D% (Q@JP(D% (Q@JP(D% (Q@JP(D% (Q@JP(D% (Q@JP(D% (Q@\I $(`nIFgVP/,$ ^@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J"'%x(`nT{baX|J EQij(힄vyqt:Hj8׃Z ]KY Y)[ OҏdO?w?#OCj _Ƃyj E+_W"E+_W"E+_W"E+_W"E+_W"E+_W"}@ָ}Gr*7pyBăl%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J Q(n.Gsi}0o^|KMի}wqfA7{;_?˓G@p ?%DZg/\D- =Qb s,Z{b+Hz ]թ0&]Kbk)=tV++^~+\ͥAWlm|'tIWCO J'OW38'NW37MWVztu?+3{vEHW$CZm/cߏNvp-Wo_~gxtw&ĖN۶?WWs7?ϻ_5k3ʋS.Juymojn]yh_vs1+l^*>.?<\x>ݹjyX0Ś>jJ:ԃṽj)՝~T7t˙aݰ:ptV ?^n_|_/L6cO}~([rT҆A8kg)Wo/Nfn.~[Kh{V\i Cc]]r;:P#ͬ23l+9EJxPPc> ]ϼZhU$_c(z]~g" K'Z0Bv}tzw6>}wHv.7^}]WU?A%" qV)ŸG@}֙%dݾ )ܐ_R_Td9 gb{yD  %I;?8I1 Gl݇?rcȝݧ4Gook@p,XD;utkKi&?<8eZA)FM&w> <{NfnE̚ҷ `[0n":*x6:=il.=蹾^+p+zX֤6Nr۴շf7׿F#:ԳYv\}y~u{՝%LE .VТ$$ z]ɂǯӍYb1kr׀ ^.jK/XFwR:27b2!/_5M bEB7('&VG%YEm]CՎtO9>1:H %DYNd'եSˊR0r\*jAUH+fE@Hb@VN2yU;*чQy5q:WV?lH}Ob*q(~MehF8z}sJv>Ϭ|(26Y!;*PiUXnEXMuW/.Hg\C묦%E[//~q/7Ӥ>ْc,+]4>xVd5U(tوQcѧ//9ҎzG ؍lxh{OJF |M[; l(#7~4J(V=r_/z·[5}}%iRbLiN-ݮ!Z>G%F뽌G(ܩF!9t hay3NKc`))ς^ˌJ}WXT74Y@JҤL#1[ 1+4Nr3㙒)^1TcGwzj6@$j)]xaKWI{jyf_*a!p69*;[/UPJ*KG5::nu=h ǣ')dBGia!d[x)/]QT<0-zU] Zso9>ހ,%!Qe9b8r6#-bTrFf,; ouZp(`PEo, !+NpwӏC 7iKnnдji ѡiJ}q6BT/Amv8mcSK ?|a 6] Vނŭ1RO]1:]sj?_cG|p@Gw!mq pmJ_}ZU= 3R$zsH~Xˬp%;D<}-|?4MiRS4=Cݳݣ{}7gwM qqwMҋs 39^K/i쀶FwwZ9.t ݟ#yM잧ŞY ̻>\#QGE><:6[s=J/7ͷ,RP虠} %q0'g݈{-Z5maS}|֝,kDQcgcܹ9u禛ʙ'}}6o2{ RiyiUN7*d V*eѪI#C :ALĸ^g#}1!"*U8 ϰBTHf \Ʋ w6N &x0n;qzEC"|Jv58*H޺pLo:~Ҙ)F`Zﭚ,_n@Ewd@Z'J,HI:T)/鹷VS~ET^deM4*Ӝ )jQ *IƊ:tPh-n=Ƴ{cqErHjLF{3 <*7rZJ]ެU2]W51<~0fzjIwq۠#0^ĵ^TQ zROGsldvq7M+.׵(uUψUW UWHI2$1dv2P ^y RW{H>F#^z-ի>{hF6BthmF&(cBΙ!JL+`qlڬ35 ~/YCiM:!=b1Sɤp(=mo}7OHR8Hf q.1!oy:o$YpAwĠLT S"CSz"hVI l +7z+u>@{AV 3ծ̮<uW7`vhkR@# w:l>@J$'X 10n%T Y{xi\*8 |ۜ3sphT9]⪉s؂w`e9|nenN#|#fZYvq-_] K%j8Q"x\Xa!$։oq=u,;ӧT^im lk}\?7arѠ9\\dI*L"C+AA"sN7g9¿*bउ }xj rR*zS3qq?#=ݠ Mqc6'jS< lM|ᆠ]%؛^K!'ҁ&b&%fPw[ClٲEg{~N ][h7?\lG6HKPchS|A⭝ pH3KrI#gqpsO~. P7&J0IDIւSkiiTIFPe(mRr)i[)E/&eb<d&rd.{!{ĩegWz0;o0cW \_ů @K*I'\eE^ΐQlɢ{goR ]ܻMGByJXgt|[ "D ,.)4rˍ]]}w_h@8gg U#rnG|9~R[?q=p71m!^ tYGfSzP{EkȽ+p96?!ƶ}wWlfp\}]v05zF^s>hVW7?G| [oz`.)kOTQ5WvsQ5͘whraƝ7ic+ ]wQ`{)Vt֊ʌ~kjt8CW֏Q8EN@/8WR%CpeY|׏XYc5ˊ.D2jqmEQbU2xͣs);W"J#Ÿ.EJ1dtYKֆ̽77Pk&ZdC! RfÿOvdïQȈ߈K1ny&S7KH(R^ sX$2L,Z,QBF=ގ@6ejh ӃAhoH)OLp6df!Is^`Q"A/fҔw[}[֯jw^M/7WFO Kk@⚁& i`eC`ZPM$p9v}D]Ӈ"Gҡ`?QS8t?oqW(;[ݬ.>ee])q kXYbIi6(Fz&jΕz~ߊ 髛tE2Λz.{Ʈ=)th)vf{aW/zz/weH$e3HhVL02Ս0xZpKr] n[e+mK4P.K$d0E0uPEŖm1T0*5eb LlK#gWvQjm¼ c)gy_#ߗ5džb^ %GLQ$'6k\Ƽ>쀮I3Ƽ^oVR= XPBuLyU %C9 073Ay=m뗉HrtônqyTd2fq&PT2zRH8MX"XscK_kl۳L&3Y kDRT0d6V`fLHIO TYZ[Ac$0]~/֝Y= ibA Y?HcJ/*"v\N- -<S#biEIkLgpF}Bjv!=%"҂3[a}t:g33ϒC B@ ,Y*$PÕ9 4].|э ?Ϫqgng)'&uRSOtt)h=aYwZ9~#uv:"]'jkdat|~#gTfP(qOnɥ!'n4J=z2읤%g $rU3(MޥJ@IɵL9+['\EYcݱr8BkuZݸZ%iaiZ_-Nǽ\vv*M58BsM]?nq# /0K}^iZ/5X9ySje`Ԃnj2.v8Roǵ=|ptSK6$zᛛzc3mLA-Ji@Au6УE^gs@acnouMn.:*F4[#a0oG䪈Svª{_+tK<뤲N +d?8ǟ>]t yw G`B؛Eb{p鿽MfM #witwi4 [>2>oVf;a H/h)gZ>(-]5N\E\cg ٸ{QUT.UZ^Eb4#A}&C/Gh`7#p2q&JMZL2)@ ɔ %GDžH/ml(SvAyH4sgљg:PrW5 ?O=>d-:]xmUv$Cm]?[nI9]-4GRVR$扡`Dp@\_ɪC'=֒4}Sˉg5-\rxIZKAnh\p)۠*8st9082fÛ5¼w!)tOtإ'#̛dmI4:H-f`[M4545CZ($[Hp :IH$[UZHIKd9r+-y2b d^hM3G+23b2,ppc@)_e"E)_3A~V eaT ~?fc>YQV?*'#Ldx-\q$:K. = f-.n[w[0opՂG$@MB" *Tj݊7(#W`mF\r\\s/ } #1)Hǃ\XUVCW%F+]\S'TrtBZQ5,G+G-qT8,qWЊN=+?qUXRk>tqUWoP\7~L f*v, *T*ۊ7(EWHhU!WcWZ-]\*lWR)i ]!ꈔB"Zxn躿*T._+Ź4P/rgҞp&(2F#Ӆ` ޻b{i<qrwH;'ӄ($H)$M7/ +%Zљw;cҩk{9 vYE[H[b7D򤉑2AP6([s CB +Ԃ9tpW4M\,KZaMQ+''NX$ pҩvlugt5YAj)Bx"b]̤D0w9.CrkmM향u դReq{$4SPJą&[>,Xjc`Hn2Rl&{3% `b. L*8ll7995*pr98;X`7Y?;c6uй-zEn[:xd5Kϟ-D_'嶁(u9 (B;̨yCdA\Ieuwꨁ*e)XT#*'Qn^d0 K9OZr+xk#PE 2d*tL!Hj{5Fт>2.Җ<XN©"kAHϽJY$[tJ?-ek=%f(i`-J%mJP?BMb^_UzYF:_tl- viyl(|IA)eIxeLsf#E,^F7BM`{:}7C7wK2NJ1ws]Xrq_ZTueԯX왟qOſ#œZ}}]ȦͯsyXߛ}[3QÌw@gh4n{|0+hX0f(yOZyycA}lI1m <'+-lls^MX7eBWz޼t=m"esG),g⹗ x-k](ÅBqx-w%02X8 V &X)sET@3)ºĝXK ToM)zɄy_r&3Q2YAl66FΎv@ł}.pjw@S*'\ę5쐬jOV@ H7jH2љQQ(&)BY2JdɤZ "D^b74rf}mםݤ|t'y5=Oï7&6NA6vtϵV5dZWPo6):^+$v;Թ%7E֎U/-C:}3[I̋BqE^urw_Ec| _JIpGT:/GK,Iܺ{ۆ^͋W>CK%Ci\vsǢϻ< 䇙~eyN+K ϊ>,tv (o_[pIHmZ^IӸ-Qr A|7  7am3d@q$cɂ&8U@gYU|kmkm텥XVB! @7RT2x9Ա >+ Y\֥HR,dAJ8 1&d\"mJh"+KPmA#Rl./ /~˦//'gG 1_ brmbz!FS5X CSmŷzU7$hiDy489D&9"Jgښȍ_[;mH$9(2%YD' 2+":P'XNh'XzOAϼLZ _/!T"bN]1RFY h:, /0,ŮҼ@cmL- l-DU}8 +VD:KeԪD[Mq%^y]OSv軅W (sՎ[>%;ԧkIN>٬|f}SvGu0m ,}\m;2^)}Mlf~6R݆r{@{Bn_ekAU{trd췊v_؟U ]7ŹjšV$'w~8}Ŷ|Jld Ka/[e%@JFx`brt,6r%9rq>p ?ԸP79y㤶`0O^|+PFAVꄖBD]UCG(6$o?-W?_Жj%;qs -ԓ/ZrR$` X]dIAhTB֋8 4.lBm2H$e B>!IxK˒8l:[5= Dž]y&iɈE kuڋU=<Ć*zhuv2JRZFf]Tր2Qy Ƙ #O#\ Oi &y/1=IcI߹dCdt )6 ޣASIM ]tXseQZ7lB206f 3tP[R,y%ې)$02Hɒ8Q5VN"+sQ!x_^K z=;[ru+Uhܞyeuг Vpm@f 6g+@7m{|>1A4or >u e ~o1aZhl~+yZF6pYiٵ0.Vv&cbԥT' NEpw+EzzmuC[H2lTJ JaoÃ`|gZhv$oێ!ݵƻ)K xώ. /ڽ=+>* GS'ptcኩ}\1u6\1 Ws1އc?h:b~[E`I+ APWytda7xg{P=pJqp0oaPdFkFѥ 0GƮ1ho%]PMk(Q,z=Ci4Pz &ǐ2D096B(A6%Fs|Hĩ;G/rT\;mYmqı/ 3'98 #wB֥3WA[S\ŊVJtl̵ꢍK^ .d|,Bơ8Jr(Z?$YM6DɮΔ hfK3j4bȩmWRCI$o\vtr/:O*<` K5IzD9 5k`&S#:K1~K޽e{Kre{,).+Tgy'cS'MʃSq^Ş.MhSV֭ğE= DP\վ,3eփBV^&PBFNu 9he`Q% cR%̃"U]JJJT!(B^h ;v\4()y*%Kʘ6L*ˈ?oOBVq:JMfu,,_/"'пG>迶?.Z\BoDouReS0 nu׷fh[>:q]q)pܹ'khCT?çw&y2ploGN&?w,~;Au#Yv$0tn~s[X~a4?ez7cG 7g];G]QݫAFi#u`+4rRƩ|#O.aNEAleF]8_~xOR_~W?⎵SϞߞ5#1˻0n54Z{fh|q}[^3WBո=jH_O~Ih:'bXV.;uf՜P5$/?f5?'D??ߟF<&ī_teGKa'<_Wc݁zR|&~wF S=sI:4 AM,Q1饃 K!'EXfkrZ)ɢN&J2XR6D4ɒ(*z]ÞN=)>] 5J384j w6HXnnf_$M2u^X;]*vV 簯) ; o ֠%( <J@PBj9@ж4m6Ⱂ^vF [-{M/nN>y^dT)}GNUȈ#,KJ.J}u<_ɨ3L/>PBrK$uC MyMdSDZFW:6ʊ}.u*$]+5K&3&B1Wi$E"k'/QN uSH$U9W.YƒbБ'i]R(5ΖY4MQ,D yNZ˜H He2&Rm2 q?D|0Omg P*BZրL_Y_e %)mʁM tWTpӡ[JW]b7;>TfrS{%ea7ovu݃5lZ}})}Y&g=m!j%Rݭ.E{>Ǐ+Ub3g4yց b^%ƵP+ix+^SVy߶\RD^|.Z*6=ʑ/lՑ].vHͦel4$cW[m! pp5zoO3^^,1l<]'tqj.& dPM.t=_vUx0[Ǯ5nE$xe&eJ!Td}Sg3|Aʜh[IǎvG {ޓLت_՚kV ~ȝ-:yEEܐZُ$Op4m$thH:wm$yVwF4O־a G[HVعWKAcG@œr^qK!&oܾzX3hZQ[ ޽*)B@Yhj=E%Q@ *PIgk렠T)S,:A"EI,, cqtIJ;^H\!5Ի1 STGMWNW^}?x걄=˅= 2F،{1 HEe y+ K );흶*,^,Ba0 `(aQ`; 1JeD&It%Y(0! SAm3 _"Ii[̚b>Z'vM r_|,̈v^h F2ֈ#M %!O` 6 _IKJ\  Osq5;˃YkzYIIEFBacIb.z="T)2Tw={dLmԁ[-7o IO )T :+sECD_Ͼ y7l}@J+]bnp3ؤsm_!%vMC֞y:F_gKVm9Iqz$Qꦚ*u-vD}}.c΋\ GƐ8w݇'A֫z5PjpJ/QHmsW#elijvTz۔*Mpvz+i>x1sC$xE{uε-n(~ZP_\ޮY]Q>ղ}=]UwR}RG{dG갽/nJ㿬C?W{9LϷiIWJW~>VL|fodd|Vjiu^yVcw=ٯl o__?"Et(IkMH<\ޡ88}M^nV7K>ެGSʞ;C_U,@Mm\+o܇n@_U_Y5N7V$4M@S#:?(o>~4/y8R>*pu~PH){bгDPS]HϜ. w<1n톪\3UλKc6iZ_?ECI n;Nt 1F9v{mv'׏?x>])^=:F^WeF9S:C]BV$8 Uit:M3lŐ}}+Fzm6«'ñlEmiUX TPT&jJY9RK*^Ll闪D (Ӵ8X6upʡcQ>n\">@Oz?v`}Il~mw?w6 X euF *^,gڇlX;wKpӑ,OYgwj(PYg&IAc*Ȣ4y!*V""%oP r [b^Qd,+)FT`JɑCpzpa7\/C\WQ^gFIKiE#>WCM5W.>sYK4kjdzY.c߬} R9 5ES9>#TF|Ey]K * j t᱄"/6jeQhXTeS!h8=AB)n=J|O/:ZPN֢S*yi~$uXO)A&}y[P7fusb=uزq(oRJW|Y xЏʻ&Tu3Q:4Ov(Wmֿ_iR˪(EmT7Q3)hĔ;mJ7M(rB oltX@}.yb>E~SӿLvo?=8w(YG&tDӁN)+":y,9qW6s^K缾s^]Ϫ'7;\Wvjiu54D R+7@W.ت{i"ҕtU<"\|,"ZM"J߹H!j>R^銀htŸ^Ţ+uR+6jCULbcq1 т)!j(mDb` P i +tIWsԕQZOufwcV>@kYΛ-6Oo:a!̙gB/ ?WfъJ*g7&23#@[m|9h)־XWUm̯ӿ~>o_T/.VW﬩׫*nAͼ) f_]흌Y!ӛ^"sםr؞[lW/:Mm{1^i87\f&R`VFZ XpDދ2׈ UF *,]IWVJunzhcqEWD]WLy&]GWJ wGh+6jk!#FW c2УʺtE 2FWLGWDiCt"2|* uq7^(y`T>{HA9 XKFӌe,fZSr6'ތrႚ^ܜI]W٦|xTzCxGO*T2cSc,X(!\_4^$59[&/m]+ ]L]^gUg7fH^{ᾒrFE5VhQe d*75*ͳ*\wJ칱Y]Xp؈Lk]d0&̛TUΩA1ьQKi!1 Sڴ jcAL"`c\FWN_ɔ&|QW^{Ha6EYŴCQ. A8StEģ+uѼbZuE(]իz\ L j4S0 יiu5Va:p0ت*"]10htEFB,bS|]1IW3ԕ2JHW׮]1.D+ˬϤ[ҏ-|muC~A \эVtGyAˀk%Eh >()C;)E/eXktEF3øƢ+0%b u,JIWFW0]RGsԕwZ{"] 67dLi+tw֏+"Gkd,bکnFiuJRU~n)ډ/ ǝdNN:-UO]?tuLK qEWL ٤JykHW lm4b\6)N4 +]Db1]RˤYtn ^us\]q#PS#XQ"-fuӊ?Ww(mc?^˫fŏWP}뺤"2o]u-1V^NT Y^賵~xO]RZ_9<;wj;;gݏ0\gsں+yqqӿE{y|ёvvNîf ̯{+{um7GJS}pQ]fAb19[ *9~O7(zV;'Jswrc2|eWUcYj|Y>^˝3j=~nף Ja3au%Pq9oxX|=>;5V!}5zYn,{fFG5m{;g,\Ք`EA0@-/?jnK O5 lOZP{hrySDn/7nxm3[+pf7PA}QRXSI_KXܶڳE'^Po~u'BMT"7Co~\~oFh6n%6x}>oG_w̖1߲o}{jR(ͧՈyBӴoה{MѪs z'zJ=xUzF=B}gtIz]Zz:ɝ)d@휈hIk\,(Q5Ms\$hMgb7^o~|2 ȬndE)hRUڋ7bx>lߗ}֖Śeo|c/:RQUT%G~^3ffzm z 9Õo&PSJTɧϧok.0bڟbx#zb\ibzb)Niasz(c)g`Zh1ӡ)})B؈tG+•"[+]QW([_0tŸƢ+5”.Vzt%{V\ L `/ U/Fk4F]ɤc^JE+)K^N}0ZuŔV$]PWJjDHWJD]1s)Q']PWZAuE8qĢ+&t]1egj|tZZ`FF+h}8m ueQ; @)D@e/c8'f@"4; EL1i Oo-vİtJ'][[0>"]2]1XtŴ:x]1%Jy#KqEWL;}R(M4*p1EWlT<htŴS].<2^DWFXE4jq1]C)KrsG7e= {u`7? >0ʃPztl >.8pbje}t5P Z ]E2U$LX ] BW-c6'IWP1,)CW /} (;y ]I r|Wr (o/*#J11WRzv.Ҷ>z^zu?!AP%:Y='?<6L B~_:;۔M \~A75wW%@z1^f7C/bG~tħ{|r&m?--)}K0C]|kd5qY_Wj}6_@ׯ]~ayDs*  OOWGLDg5x#Rޱ>bX!N`9x?oӻ%헗r^5`'|:v'9d(PCo[jv9*7qw!iWt_']ow$ߍ+N|5\Ok=\JRQwW#гn*TP`ˬbKQ+ZG}~GS&U5f1:kU:eذO9WS5{U]*SvnTt>>ُ [Wj%ߺ&XGx|;N&uϚՂ1'„IL 7'ZZj j 9Z*\ #j QS)9;Dݟh ͛7o!bwRwVZ+Otnf()n"7M)cDN",=vX Bud2vcF4Ccv..bnl.HSZ䔔ջg@#5"<ĐuhQJ=@ܽcwd,TD2;  !FVM{'cv\ߨ>:D!;f]F8H$7綽8KU!m [=:fDyKƜ؅),O/Ss.ޟ7'!UE{r'gj*uϬs+V[u΁z9Ku>Ü|XMѵ5b',"reߠ'̵1| 46d# V\j() AFUK!Pj36.WP4 bKT)9$rM NԭhMɂ1!Kut:ZnPB B6F}m.5뎺0Qɣ^fHh& L9'XEA)X^ TTlPtA[w Z y9xC9ۦG`CQY?k(QJ[e<*J|X`,S. kBh#ѕ<P\ZB64&cwX5@5m0oT-4&*(JƬC5M=C*D@ %ջV(V1*RO)x_b ~Ns6Ajcƚ #  ڄW'#I9G!htPk@oJ'm4dLBX Dd6~9;~|wu4؇\t詌&#)V XB;KS<=m=Ŭ}(Tʗw}CPGm玾a )X1z*D:Zζ$v ԁ>ex jw , "Ϫ90V,Z{&Fσ'rDIȚ [p,AJ35VEٕ ?AjD';6b]f ᚛l%֓DyŲ*aRW(mQARnF3 u<]A lU?dNԳf+ZOQ!,6\b%ɀn ~-Pq\,*wڝ. q@D!C1+Pv# -(`jI@BvBѸb!h9H0H5[Ff/nla˳Sk=0f/FZ,WМ[0(k& :ݫqNﯞ]V|o[mJy{ r:l.lכϫ^*`}Zt`Aƛlo=NV^oW'/_j_('x?mﴺ=v7s~cOn?ۺMc;|uϻ]N1XϷƏmI3Mֆ'3\`(''F^h^GJ+ŧ e;H@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 tN BN ~ . B@{!. 4B@vQ}\N ~ h'|%N q}-'?ҵO hN8s}'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@zN ː"Kr q.8;Eq=E'[^@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $NbTB$'_p=/ QH@O XrqH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8胭o͏g?cԻÆq~lޞo?tvIƥv1.1.-+1.=[ `HՀkRjest5P ]=I2 +,\Z ] ^;] ]=A!-n\ ] >o()EuGCRjGOWe BWODun.gqe޸u =Y֕UOntJ eWg_|u3,5x^){ĺ:eͫ?nVȿݞvᾺJ5o~f\:G,Uڜ#Aqwn j5ڜǘOz!_a C>k'];/npϧxt>>3xqo4Lhkm\`V6ߏl:h1E3(J#y-~%Yc++5ִL$H~uK"fb(!a+Xz;cs)iNpfM2ZFO(%糴2DWXUPS*'CPv J!̐l]9k8juhs؟JҮ#]iĘj9%\*BW? (oꛡ+a׋.Brqt 0{vp7Cd]mWB]-JtkcR +,n ]J4|>U@IuKWgHW$ "" pk ]:]JZWgIWsY DпCUc]s:]Ttutpc UKCW-#NW%-]!]0D78Qc* B:]Γ/8| QQEaS3CA)r VҘ9ءJkבּw~c_`BL:>u[g MAhl(aż;F AF0SaN2Dy?M==F[Cv.8Kmm%jMe&3\^=*=?޽ {~c_Ȝ3i2zL:uI)t;{V1vR?8T8M%1b) iIF6,V !ZaAL$bK"\Jl@k2 h4F({EMsבOB:M[r43P:eLs6q\Q|b;$xrkEd. 3 J M=OcT1^ux, +|q1@ O4L*"!.2L̳Xˆ{AyOܶ1`Ϗ TOhc֍D1,ȽL1-IC{;AUv8H!x[?jd UB21"N:ivaD(NNz\2hIl];;2[O}%%hcg6F!J:IH0Xq0{ ܇r(a'6+Lb;xt Ǥl> j~oe~y)YckaR0F{W~j`\k\#|PeunAONREP%n^ ijwɏ0šON.V0 2ptQF J1]~t):*Vy@cxE̥ERp5gph..'\ i~O|(X,ѻ_:&#,Eme,tyX.Ei"$k7JE_R*I=5=q OfTaCgNEN<,ogU0[Fs34dпۖEVG:Y _ٸ?z+)V$ח4+em1Uay8'hЊAG/Ne#HԖJ+YWj]]9-daHC$;B:K}gtmpw2ѿV3eLyfQc :~sow?^~u}w?7h~ {X)WlQ4=TP޼h&ECֺhMUrM+Ç^Vn6Rf_f_Ԏ~ѲКޏ[ ;n a>,/" QIg.E> z x.dWg(xT2 zvGGVs$ϑ0q@O0%&EKK8r3*X0j;Gzņ9Ez /p.&n-t 'I5EXXsCR)Ď;cy,Q&B+cB-WSxנN1$TsH {[L6 f_Ql^IXo/Yz]9rU?cly&y> =J[#F*+?/3J<$/QͻbY!l29C";KF_0N 娳&J8rXŵ6ٶUw0vxn{[_q x2B}':n3tA;~w>v:_C#g<>doݍƟF5I#OXŭ؈tO?]2*s/)w:> {~WV Teg,'L]2UIuJbLgl|d\O'&ȎsolgePbotWd\ ~fiiov+3IAoț .$%W)|)w$2A{uoq}~7f\4^Cl󞁌8, _F!k~d ͳjxI"rJ0}C7e2t+Jv|J%?{:*)z5sw-k5 "Ӏ$jyPprU]q\lv\RQB{DO!~V[[CEL|7ZAU$ߎQMMpvWdHCyhwRJ70kQ h5ǨR|7TV4aj27:j8 `ɷpv\ɒ` [H~Nq4 ‰z-ՋRBGf)XK0Zlv\R@qdqa ?+Ԅ)&n-vRlDXU#T,Ë4%v%BEN[$!!>iB5o,`dwQ<@ZsOzu/PIiZ\8iIԌ^&miъY**({(5 M0G xiUvJY;mfmrî]DߺTt\Mqj0!;*v(m-Jtkc lD5.)thANWW[:"1%DWs p%o ]ZԩU@yKWgHW"5\-BWVrtPֺ:Gb DW⍡++i ]'%kS$b8U.US.>t>/;Weݲҵwa%Փ;u&iAksoxIUK.*֤zuzA4 )k$8)4*}4 (-M#Ms+vlqnf5p'> xЅg89㑯;uO?9c#c˸nܝB߬TLssB1EjeraA:]~+\ۼ\H7*F췐?}')np1K"fb(!ĖXz;s .5U Ҷ+0@m\֘5VS׶F=Cm+F60V1tp nN@!;KRIդ%ghsv\BWso?OWe54@KWgCWZ(J4 5n ]\iS*e';oԆ](D;S|쵫#[WۡGP+ծ]%50!1t2 h}K0tzA7 ޟ&vʼY=kYMFQy>1VxOtmnF/Hh҇ڪf+/딋6Pe,Y@$5hT%&gOt8 5s_3#c6U)IDǩwYq*Vv٢Ŭ6-4ghv}򍒻BQ;ɬ`c +5b˗xrq h#* pۻJkج4Pbffz2OFCW 7*)tEh A ]1\ XjteU V]1b&_]1Zs+F[BWBWvŀ`FBWU̝. hC8xd\@㣥 r[32c M#R%LwEmAA^J֨ܽ F x2&$b^ ]1r+Bi.t5@ F{Obz'm~t(*g%g씜2 A ]1Z̞e(K_ ]iMZ(}* p D.Ih4ENDڕVjsm^ +{D>wbXmDW ڹ4b=CXg$"Nb^whcDԅHW$]w\"ܞNDBtemP ]9rs{tp쑳!hĮ%w픏QK`fNIaiF.wfbin"97[J,wG4ͯWٗ$x@l֥2(Ii !,E6X`%Oٽº ' %fO($ 1F6{%nOGhѹ̬A硫@V@b:-m~P ] U]$m2\R[8 eT%+Q@ {$D4OD6ĵN+]ꩪFAtE}@ipBWQWjteΡ$& ]1ܾs+FYjt`õF ]1Z ^ +;퐮nϭblb ] of֟ #4pܣ%| rg`KϒbQe pmP^bpm{ hG%0Rb"F+F HW>:$1kF--엜.k8CӒvpOpQwhNWBW_ ]7Y`4絫}zUJY1I+Sꩪi-.bs+F[Qg+|DW8*+1R+F !]20!CWW[%-Bt(.t5@(ub֮ʙAJ+w4NWBWC+&vRzsGD>O*AnM+,cI#D,U'YwH2H:&&;Ifbr:_$o%*bfT"h@/(6#V+15b muMy0z1tpBW6fD(*t5D B;+Q ]1(& *h̝%4!UhEyW X]1 f:=]1J]6eA׮?Gɗ* ٻJC{LC2+H+(tTkZ]qߥQ ]1ھMC--t526zg8{Wtp1H+Bɝ)t5Dgճ-]`õ( ] J]1ܾ 6wbHW&-;RCD hMLĉi3ڙ)%[PE2O,9\Vh!e0Ve vp LNPpQK+FSc p bF9ޕǨLGBU prABW֩Qb"]EzUI 8h1tpt21J[6UoTP9L؛Q׮ҟ~* l&Е-tT`ՂޏOw4NWmҕ p)}= n^ ]1Z D]jtQ[vJ]1\Rb %"]hm `gbv. eTHWLQte<2 Xkvh1J.kВ$NJG۱+x&3If˺wGmRcm6dn(JμAKDIyQJlhoByB~G%g)dat(+t5@ g bҠm_$̭5yg+TI+ ^ n:+FY++Q@im4oJ 絫4SqJڕK+Wꩪں HHЕnH+B͝HWF{$~ +"&b ] )]1ghѰ+b6H+F;]1JHW֘ bVE#mȞ[?BWCWEW9tEpni()t5@rbʪ7p軗/ǣMuxrq__^]TzH{J{\M+-`%BOl lwAj8p~痄tBVЈ<-V˳r/z\a뻦"W^T+,yM3;Guq]\ZqY6 VzF8bruf^ck/w0t>24xvzr'G˧!̀}8j~ym]?~}nVny^.O/'FtmBh :6X*EL0_Tu݊B|^-7#Uo^4X~WhZ-߭wH|kx I]H/(vn:|ѵڐ= Ō<؋dB܏ Fmqu񖃉j¡e*^)a%/6''{KlDA^ʼzysX5c^{LI.[K,GJcrgW){=$''߳9c(Cx5xDd?t@>|d Jӟ-'wﴕ 6?kլ}ւ"c}̂Ztp0o ĪPxPNd ȇ0MH+ا{Bj(RƔVI]dK'/#uuAGclOYU^u1ݍGF7~ cBE:(^^KTUB]*=kҙ7jtXg ͺ[5}7?X$t+2ȞCOA}^4 ݌*\ẃ\[Ԑ>4T]((ժލ]Ijh晕Q.<?ꠝ_oݡYmwMzԜ;d=o6j\L 8j&jU\@=9_>,.QX>&NZڈNaaVO4b'hKCgЇ!04浛CeP`v5uM5_&k*k%fr֧-[ck$<*>2z:#љl ꚕyk艜[Z=3,`ṴTmjh6崶IRPuo sXG4㡚5 |GPUb4|[_&225o}S}0.wzcdD`lҀ?餦Ȥ^0j[ HR.ll)$پ1bBjo&qJf7f9å~=YXM.3b ^,Vbd%]c?t;zۘMy~{gLeugh?#I'tUvba߾?Os'LOWbhz 9{jf&<Б+ .NDO?I^%駩3#1+G$6D1Rh`K]O'5i;"*%fKr/,fn:dDfbbB4%#%KYEW v-NBս~[i]~iOp^t9hAG3굼:zDr{6]7z p;Klu>zkN %7vXˎL<˭͉_rw:櫳5=^>2Qj0|DBHAVK,?I#}f/=e|tݛmlsme?j3ӺQߎB{OYȞ0Pȕ7Đt4_`N hv|$ |k ~'e^k?uP fs;b-|(r LCvl̔l>>k,:kpI؈R 7ree&)Fuh`+͠$z6IJXJI\ݹ5ي?iAg#]yqyͽd Js+Vַ`(~Q+ .7溞\LXlbd o:š$z/Inyx$᎕, *cڴD8IV%S4))y/tV6yr,wvVBa >k >wJM:rS~;A]h å (WYg0#j)Eбv$NR(MPr$k@JQh*E)tVzt,UM9Po?EYɥ>R:Vm1 YiH&I.J1IJQh *C1>kJQޙ7Y;V<x|+x(NķjڭOƅǔ9m IҬ~wٱRܦc(l*EA;v,<0hvw t{S.1.Ye3Pm~<>0Y҇|O5; ӷzROg\:VVMLmaX+Pm}@OF.4>GrA0ݜI0pA7vHhkjN:@Zk:m0Ug&Nkmqjj"N87!Ndb9mҲnsqKw _o=vfĜybcBM )f $z ĩqn`+knGEڸޮڈy݉Y UrIJ E]6 kg+% =""yE8ZyNJw%-Vn~QN1#e_H J=4 G71C?>zL룎ɨFBfuH.U\96/˾YӚ'M*LL4>$:_Vyp`36]eX\1FstS ٗKM%ԴzsXՎe)o>c;9g%T!W_z ϊ/J~+9\Á X9Hl싓㇬5t(M+IU/zH8VW!g>;ʌr p{%>uR :8xX[%pJm NI>Ơ<4vHUX23޸JL,3왽t˙I5Xv68ֿReY ujj;V^%Ysa/2;?"K (u^i$9{.G\^Jqs]%?Nܔ& -U@3%f,1.Q[C ѩeR vIVKKng9ZfL&}<9Y7G2*XNxyo Q|&?MXT_2&FΠf+ubCtpJ@Ý^ Ş*bNbw|1 `Y{z!^5aĢ{>g`A79j5+`&>-5*4l_]cΏl:e7_%{h\\],Ufn43EHMW!#XShqQ"HޡуJCt;ѧEigϏLa"dT%qcmJ56̨RݿgYfCCˁtw@V)+8RO. 4t@ˡ׵V%ж2T^ه\YllV꯳ɝ|?.nybӧi|r<ܾi:r ^.I̟[*\f_JQ^_ïR{MRB "Zt>NRIH?L^X&k?s}$Gޱ^ ?}>Cq9?V6IVMm-D|0FkĆa1Qg שH^UjG=A*zOnO{0ʁ$:K.`BTpIrR T/x2_~1X򡤼cXl`Ldi0e&HsKqVSS?7jLJ=8h-c6_F (/01.?*~J# |>ۃFk%%^ /W|M&XHr F+фk{x~],&$GJT&pE92-Lu ex$!gI;7G=V "M2c'y x[`Jxٳ6)NnԞV`a=X@krz[E–qل ۴l7o<[w3  :㭭!l0L*۴l9IH{boJ9!uo.%kճf$}Ysރ8H:Fөѐ/ 8녆Irpt:Mnı/hƴ'b!|Q|ֆLdYUAdznĮS `!*[ Pb$MׄI%/ ]pNe^R*e`bbYdAIEi0$lQ6wX EaLMS)S1i>ɑw=->ʁc+3}xm  :dXXkh9EocJBrEd4F*TYmO#XB3RtxAthdQY:i }xϝOf5snmyU0$r;< Df \rpuX^r[(lֲa}@X<81OMN  I3'g9y4^zI:xgmx^!t`SÈ gqki[Ejt/2L;  c9C`l<Dž)БwC!`i1^'D.wI羭#)b $aۢksu͝|?u9^fd,k c4-̾!#Tٳk\KRPՐ{oVͺ$zD:ScFL4C? ZVTCX@ ŪP:Vէm7" 2sb]yAVW,!A #N!= ^S)EL=094>1""6yb `ZgX<'%+7¸-Y2!B[*oi |1;l`A,Ռ3di'f-P- [iiC.q.IcZxY,$$s9$aڐ+z7~H2\9 ,X!qE6eEҸ 4;fP, 9*/)hCސǏ"#SܓMB1S&~>WG3{Fn,E藝 ۼ_ a; &lՖd[/ l*%JnYIe7yfL)h0<F9i w8݇2{᷿r='M)6Nكha,S=' #%VG:>QvWG 6iq@CNb:I4E@H%#4#N`pG2PLRhhE.IR=TrZR+BF O>GҔR =e B%};-uhul=}\Qf7l&{>ŇZ ~iv!=栶'''!1̞kWB۾IY1vڳyؔAItXiSĘ"=9w%J v&ٵ-GB4xZĈTm3yRN_~D8رLBFp)T#Hv(ꘪ$8Y m~TDgWL~+1XT9ntA؝vKனqSsa4~PeԓЂ>^V VF'D•C|5*:.e_AKkoˢJPkabC@}#EGTi(| X-ﯣA+-4I% ؓfTsn $JL.HU('$ jݴ ɼ'f3Ex:OP Ԛ؆VVnƍ"1fӢZ$Gǘb:%EcQ&Yࣝl.wk[$Kq4S'&vǀɓBh+ѯQn g/絉 '/?@Bk1F5 m9SV rJ>^+s-e\qB%1SZWG-'n2 y[R F!n]uW÷ylK4}*~jJ:T"zJ4Z%,IAVj /!!6*ae39:* 2c+lLm]26bk(E`$anV>i`KZƢkWTkw8X)BvS|:Q cfmH:L"lA+4zdWd⎊^@t6>W!DMUϸ\yP)T_y@2%v^|SV'K4+r K)wܞb2oW),j>[^ӷ(GE|Xs\~5_|jj;^"]\VXb,Xpkf` #cg_dY1N遼SPpfYvt t<luxa7 kz0n8? ]4/3# RNgbT%iO{_GlVQE]f6bzQi1-߇-o[lQ!u[8̢.`.mi0TtCؔ'̈.b4U+qM}otȬL݌?NF|4oz?/;O>M?~9FV)J)G XQ[ );i*fv8~a>p]O 5 ѠfZ B_ŏOIyO?kmv:ۦW-G_sZf2B<+pǘET6cNp:{a1y yKnai82d0.Ɨop.p!~Ѕ>wf׹OIUA]{zVGn ƲQJ#}ܰbŒPT26Pj+ʌ.\D/a(}. vهqK!qD7h>ԧƱau1u<В鮧t{è01 |Kki 8H6A8%%rL.j`o_s#ݦɘN]W"5sefyiVKV:l9! ,^\EguEZ(,ì?JͧJµ,/JaT|\#-FN,n|è˩!% ض($G:'hkbR`CgTs1Z0+1`%9d}ccW*I ylEBA#G*uDq2> ]Q@;# G×W>_TjԂh DaͅJe%CF;L!]{Kӣ6fT.{P=OaIvAa =߶%+/9$Ǯ]Q$l&2챱]YT!`69:4I"Zjv-Z0E 8S),k̈́x!@W0^>Xx3Wii!6| d+(Z*eu>E >uBuPn9ΰB0ZfHDR(!h;'es3Ī:8@~1kbfKi6701X0e&=m4f U0PB(2D(ՖM  E)Ik }zT̵9 oUM3I4Odh78$[>;__+OEӱ'stmy{4 '#[_)2 0\є7ʝ>ܔyM y[xM*`b3Ͱ٫ݥxiGJh2ο$6%yjg. k5ԍ@PK> &E1^u>x' ͵ޟ+fESPMw٣_\Ld`kVwr46 r~B[qo!c4%ۣ@M[TBQƁROm3􉮥^ZC8")6H&42p6`(*[l0n9|nh;Lɧl[\U:EK"fwP|ݷX\4{rTٱsPw qn>J}$0H 26j?({CdDc x}aB^rmUo63f"߀di[((3 {1Zl <ܧHlneO?u#C"ě I峮H8[_?ZoKB-HѡCגܓK O'C?snFp"T"L3#W0L~?~4'"GV+2Y Z6g}:X|Qվ8rwd^m.`5O|:Ȗ/}ZwLK~̧tF٧x<w `p9ˉ5~T>`]1p~i, 7^"~VvCy lST8mFa(@dToZA/uݭoA"ϭ3FwtU@&匼>(`fdsGĬI?8ÞgZ:*cĨ! I8z<eG)ǒ]iBDW49oti(XاIgEtd9r(/cMDY-@8qїWY/\R[ܵ0g҅%]=đN#OPV_q-p؋)S"Ō̗aoFx(z{P%wR9C)!뉻wl=YM;"gpL )C܂!F BJA{^I8V](2/AFFqꬢX|X}ȱwPYzܔ2%g<CK.e.tuG2 ̀9!͍FV% ,zjXz2jܙCX:C{g [r,덖TӔ;tdW vr3o8qK)*@CKN9TZƜވ4i5Q`Õl~{_C 3Pq]c^\'N3w[8pΜ(r2V 2"\Lg3iul8%dModI(hF{MF٤x:f|X)Vię1 EY97`B|6d}.<u?b[i;l!GX҅2b8_X($ 0L)1FK#e浘9gŪ D!4#˅|#2{#K J x>Z}cPچV6Ž!d"{$Rk I$reS%k[hVWC܁04pLz.+qHЋ^tɀ{mx$샱a!bH6Y$,ȩ HbO+#qT8JvV~TY3uY| 6)I_‚Ek/FIu.‘Uףk19|ZJㆍ`SE=pjk2co-A-sƽ3^J5/1Eͻey mL]e (er < ­oe8ȉPNE%4% pb )#x*R%zA,6 r*f5{T#Qexqscp4Ad9@)qRAV@ۍ YB8ئ}5dI鲉BY[Nۜt<yF$*d)؄B9oͼ pA>C`p-MGR3fVAE1Fy(HS>R,jΉvlPl0'e$1) փO q(NrQ |s!ٹY> C`#Fv_M5J~yZCD [v)u"uѡs|[2Aih5tP*_ԕdvT-⣪ͬ D;2(W, 'ASgtMD8-|J:RWl5'Wȹ2bwѻ6"G_*X =(!+EID'^ H&%4-,*[tJIc"Wm +@R8& I"!}Vx6F3^(zT|Ef #P-~ibT-o4+ML*\cA,-u1qb=י`6Z?SSaA dL sѴ{r8+X%4ޡp`jZܢawqctE0xаi&LVPݞmi/i %4|3a]c%A=86-h.$\rQ \"!+Zg뗯|5ޡ 0Za&a{Q4{wt),t&Fb"WB1WpMЀ%w`1Sb& cHdÉ1GFK/}2!X邙{L%*9IJK˵80 Y%4~QjHUJs)Iur gp,*䁢FrzX .p֣ݔIf h:nkZr+ttDVx%W)zz5 GQY~ AUU|Gnj %[rэ$ˡ$ɱ+Fz{OHLVIdR|)2*PO(μ_Vv&4n1{8qޱ]b9mӰ\dz _e8kZ\IcCYKτ2psg~N2? a>DŽu0RʧE4]M iy៍R@3yL$oxKIͧl$7Wʺ *sՋg2 nH8"k.lflN̹m4qCQ@phތlӎ̘3:wK6;ޫHVTKʎt*[̎-?Gq'U1dn }\KGp n#YxcV,oFR=̾HP>Lۘ דr߿.xB ?9A2ZA!QEAq8hFjN *#y7 ;F #`#hk:w14Qk)nH| sRꝵjM#NY [+&VҨ]lHs%A):g%;T=d+XCv#D]EtЖ1C uW/Zb}ueDskho|cFT ^66p/M̨_M-nnM{}5,Sy#h ϶?zh([SyG+F%-#R^WTPّa6㶻qVUyK i[4Ap|0T _iYyAzMGCubkN1:\šW== 8ޡϼ?x>MEcq΀؎ȩ`F{3[pfYi@IOT4T'}gk|`G ʡ\[6MTU;;Ÿh<%q_&AoikTRSí.fin-2.>NyuLbF.(p.Z+AŎJL*K TK;1%#IwVju-2-iq4JςO% tYfh4bŬG8:R"sp÷pd8.UJN/%c\:.xx~w4zJ%iwpՔ_uByKL$`pA0$MrTBcgL zqVFdZҁBF| [b$GNp|0lˁOf ߒEL>|Z^:zI*ny$ i|y-|!7CLYd G\d}b xF{i6mg~<<`SyΟfȆQ6[kzRM2oŐ,V zwUb-"RGitjNĪ[Xla9#ESuUfPT閏*J-ɂT4 t%6Fʖi\͘8j wV5Eb?I-#I!T"!i OB[˾1k3\/D#_||!6ߕV>U,Dy%4: Jώ{Դ\4nv.zUt/-Iz(wUl}:.stoʉ0%HkElyQfi/2:h̤A*a q("JI*<־۹I&Eƽ>䧪)wg'\?] y~9`h9#JA!cwhakqP\dHyJbӜo9(戦_QGBi]]G9QR5nU#a[jlr`TdDM딺[gI(OL; ^gD".BRM!T;ֲ#wlNqw09}]*p?6:Jj*ӿlY Q^|nӇ8pBv,dYjR"\%tEcWO;GEzQ3!p-#8W46hk!jJb0,VO0Z4^NTжSY/c\9f귉:+)iQ9Sp-O8Q66#%i TE"KYԥ42\_,|(_smHBRKw yM5:0Oax$Y+4P (;Lcw9M+(e-FJS$a:e9.Fm@yʁJRPORF4⭛ mqqND~=cCfvuیYގ{.OOd8yMI,r*" d*hRW "4oSёvYP)݌Fsup)<ݻ`?"zfA,<%ٛ{s<_~{L>Z70)?"4I ȸR;;&!/J' Y- S 댳)՘ Oxmra Z̼N4P|^#/`^[8i.s;8pOFbG&4_)#ҟCm5nHS&m<:ڟ쳲"Rjd,( Dc>?ExOMDIBo 1 w/-sNԯދg/{lc ,FvW -Hhq?}y#l[Ј9x;Otրd2g{2Ł©Q0ES8 .zW_?F#,dCzg F?x4Çm&veVF_ܝ{]ڨ[vg~O IyvV.Aec/L zǾ4܊tNj7q&aoa2Egu7F׭C;,JHD~P kA9_ZNQHR4#97+#5?9 ҡ&#jFgzqgI.Ral +L47e I8 R,ŝW,8L :B;~3-ci DJLvlv֬ZiJg#rzJht ԍA2 R!ʟ >,ERiF< [zsYBh:1*'}Ӯ#MZ`<^,ZˠfYY$Bf7;-}0 $x&!uȴ(Q&+bbKnnXQ>3=Z((_)DZ7mTl|K\gEIn(NCD "€mn `Cs(J?x2rpܜ39+S WT=^䦊LZ>vF#y9wL 7f ;eH+Og0Vsظi&5gev;bߊz&CG kUk$V@@բ;4\~j#%x[Xͼ#5lem) "MMLmei0B; A%0:M0+SL+3$&x]&)SF0`1``Xe%0M$ &ϧX[Up$zDrONO҉DG914s`F%d 3p9sd)iHK%V]/\=ul;Ԣ"i^݉HlW[~,UKZ\K#qD[N> 7io=?&NhK)C쨐Z1b!+_ꐞ+X  EdfcUU,v~; j$c$M)n̬2#zGT7*"\Y Hdz+8vݕ7W͉\+]4b|&5yR ZN n:MhWMf'(`1IK-=| rlϜ\aqI0THy~[H'LRAĊDkÑR_lwEnPb/=5?]jxvbeEid$>Hͩ$6~~,cegKTǦ(G_\iJo팔Y y `-Qd3 i wF1ϻi+[)vO\mқ.N){Qf/Nǻ0E4gQ y6;/(Z$z$#گt77"bծA(Z}9ĚMIyw%%j/ g`e˙XKMA8"~-mN7X;fTJUWa-iˌѠ2ɴtCNPnpApϦm)wxa{sG_^W$NHɤw|pjFgfs%Xo|Fs;.ZK'y^û`qS/V2Ҷ=T\>~NU1t ^bGٲU4zz&*O\M;_ *1bV(6OL\|,ycZYXڀ6'jgHm|oz ޳N~EsoX~ R@nsx'U֣Tk"fWY fy36}NN~"Rxɘc%dIEϕu*(At[xW b%0> ˝u__:shK`D -dW}tRJ`thL2'S f'>?TMt r͹Tjۑ q~ ẔlVlKCT" f":]8=-% -uy VJּ!RB{Ӯqakz,`Hy.+y6WQZ l"}lm:QjJ`tFsoS JtmyMMЯ6mEO™i?zxt<$Zs[ O+A߰ yuS۵: n8/ѬH7ѝBlʮ6mzfNCK~іĔDR]ه+qWa@R-w &5t:iu}fkN mm{fvҠI?扰†fJY,$NM)_E#(S5MߴPNO}T"%MO6U"i} !8) ~))xe |>R4-O4sm$iZp89qpJ64B)izD_bV9J"[a9'.%dAr?7Y'5Y!)Su :@ `{XH*o<)\zdVr}0W0_ y,SʶvljhQ< .iRtduz8pQӴ귙"WGϑHdh1)Bt" aȠ(ڧIM Tki߿[aΚҍU1B8Y~Y]/b7y*6,_"W3\1 Eܵƃey:td<G? .dzd8 g_0Ct ?~ͮ''ix9;[yfp)'7(8OOUu=5~>KLL3+Ri4p4#{j` [y5O_!R+a./^X :,_nysj&wgI1[~xv qLЀ3f a`YB4Yy9sg7 >oz_:TA L%4J+ ͋r4A&9..8J#[s-/"o# FEI_ ~ǻhE"=Q["AVda^p)BpE\ mS!vkdu̒䨹xK~o%Rxpkf߀4}W Ě[#?f|KGmE, #hmQ;#>#&B÷;QG딥.w$["hBȰ a-y Y4)ta_p ")0yu>z:`E 5<Ük٧r!1-hI  \+9r^CLs -@r65.oePcM_p<(00<4Zzx523k\\0>0~ -ܮ2>|8ұ LѯP$yHX ,{ 4ו(w< {(1JФGИ(wxV!N. y,64Ŕătnx, 'd'h4Jњί@֭Dn0ѺZ7&JtKjuh[}n瘈`&(5$Kwu注2Zc¿fpnMQUj!5+MwF;;s\q1Aplot. R'lJg'.#h &f7wr 3E֧%NDjcEVT<n7΅} Ooo0w$A5$%Y-glR7OJ@]Uh7wj#ZX_Q;X!0[KcмUԌ~QP^~֑` o\Ȑ\#msoT8̌6d k ao`H,8+tZ97&g \M8 ^ )"|.Iq E+Q k6Phe4R~q4b}'};R %rϧU%s} ')j7hji;~ΊG菗#To79UKۜa.ux!"sch 7߁u /O~"QΜW p?%L4:SZ,hdc9j-F*0`N)cqF\ |cByxC`>wfI7'ǴJ̥T!<:˧v62kׄԠ*.Ԙ0Քϳs>)',p{N"8G偤 *:Ul"p8B] "X2GDS-s˚{G`a,{A][=''@Nb ?A\0zH^C)A\6WL~o!=asר',apC}rBn~zYqwb,i[se"M0^Ċg$̣;MMk5wV >.J}4ޏlKNPr;\s|X1da+ =wj4Dfť^ݽ: +Gh{e90}Q]NɎI\\k^ykһ&֎.* Ye.n›>9ӟG5Eb-NeܣΝ00KJ #W{4DܦH´D_t2[Qd&h)H.iRW9mSg2N*peZ\u陹k&v zͶYnu ,^yU{m.c ;d ŌP:9ATvg`q>pHvpeԌ_ $Y=jɊq>10`%> pjlo 3ƿ0PS[cQ!(pJK^쮸!VIPNNwl=uV݅$t%&TVGYsi&_[ 89뫯:6]s%_4dU][O$ǒ+#v%;/#a>Ztڇ  4`VOTOtTS] cʲ,UEDF=<}&zF TiEhouӸm0L::r25k'\̏a66~;hbM? Hf N)z}E2؋jy|3q X$:loE E5fu! T֐Nb;hVpz;}ʹaGv.+V1$i$`?.)TF(o߀72 ı6aVZPDFķ?kJMlAJ#9{n=<v@v-Lme竫9˟x~姫??Cqfbxtuu8k&jOgWiW`_̮լy`f6xnt.2Sa[6e}ie+Luf la2h|ڛ%%сq'$,w]oooYBm5F6Q^E@zDw)NdͷPYZrzLGH.s}xڮ 04tͿ-.A$kpo|3gBEJdJJˤY+kˑ(8ۍ=ж_z=8EN"NqC9˲dk ИB9KXrK-4E@Wy?]["9o2ZTHj/M6"%tyR9`d.L*gQl ,:0(9 jv=^g/{Jkw𲨸 S%_+QkZ&Ww^wZ+I {oQzew=mOSl?N!̅D,%VRvɈVLGbr o-}BPPf`=Dǿ4:$3/E8ݖ)}囶g |H$Fnj?滅EFW`i$\&6t$RԥPM@с_p[k>sȌm,+8dYDZ+/E2WK(}#^fO%U5njԛюӣ\WSwno ??zW9ޡ"_ns`S}J5-Rc{jTP+abh5yesbx ;S4bp>Hoc*y 6{嶸Ҵ6sFԚeՌ]i'5NQTH#6lr*Jm9WQYRe.o/m߽ c̺pFKA kT #uUGX?zFM8*  (9=v`[$,zqY2j *J5C ٰɘYc/ -xQ^]ߴ;=5Ey Dd]N4#l)?k!ݦ^(C-g*:$ _AGUʉ2I\Ǹh{}v=0Ii6B~!T8vne_:Y ("-(BH 0: -(E%m}` 0{$MUD_vZ/X$wa"*2,ܘ? l\y&Uo\PƢD !U,MJooCm4ARX53(s؊XTAl4PlvbFWXL)\cQѲG;ch?Òim.Lg0X:Ҕ"4lbd@c N֔;jdv΂c]> lv-tCEg66[R"0urL0˄Z _?ۅvq֥;u 7&5t~Yݠ=cJ'xJ^ʺ&Tt腗AKb 0CQ,Edb =h9A0ZKqE˜ G#-{F ٧!m~Y"_`BV!@XC!ee|("Y\CA!Ƌpz>/J/Y93;=ƙC%|n)Q1_E vT3nYֲրAPHYDiܪҘ$_U**!F*Ơ.Jf65}HchBb[B4[Fgd5VD +^4ђY1" kGd}zglh\a,&``XH=+`C)2ճVdcba TuFN^OU빸5\rF{8Ne4Go c6:3vY<<;ãJ( ?{ldmhف%ٸ6|&BכvhgҎ;4ry{Hڛ 5UL#ZJ1]zΧsء^w+ט7ZC5MHadṚhdƊ >"FW`U0i+ (nnNu!}O{S-4/MOb`xoZ7ݜƼC RKKf;8+1oM |iw`<,M#FUZiy'K՛up/x|D/alWgMdBgLt [SVt`o; latT}دOzw+fbP<ߥ:ģZ筒)y8ݗWXo4Oљa"J,*:RrLq)]}|֓De:,h MiC4Ow>PzMc7s١2x)V@M0d|5t[Q0Y̯,DȧvUoaO9 yzgěrgUcϙw|/9.mѦ24婡QĢ2# ClήYa gJE/DЊ/s;Rk4̞"=CѪ51-U/R 霐"k‘܌Ư0<*ӫq{`]Ob` `xf,>zllcyBN?aaj lWzm(2!i>&![| Ն ! %]5Իw/ðڮBʎ,j|,0ʲIΰaȾ8dp70&^yH=&zO[qxk#z@vL` :jѸz6>4'PxIsӂSZGl|E|bq/.vG}ad>J_g82=*v<9t,^Gw\klX*ykH'с1|%pz%iX6Yh(>~*[ 4P_/ Xs`;{@˼]kmk>ZlkoE }˹5mpZg+ #K<$"Ҽvt-!g]+D/XN<@c_;(YZ5:d=y0ǮM:=9;$Vҍn5#QٴjVp[4cł@n“ʖx;g'pMY7eav8=eYEg5?ٻڞ8$W_VweKD I+hfp;j̀1{iiq`lj4]U<#yewY]^2,.m˙jEIl^K@ Xje9%yXya3}pZ?tD3$i=Q*cz`\/os Npw 1 %Hss±FR1{Vګ+_kHb\b< 8q)5?L`c)`!=3ɳ[9B ( >ULF;#;=0yo:b >(ɱNϪ5 P٘u0Ȍψ˾9En8zN6{yN34ҋ^7}J8kOa8<4sV8وS^ha1Bڷ4Boa8o%^GNXFhwx6KT*,x@Q*#Eoeѭ$[BHIȍIF(kߔT>9emoI)6#q5)-(!%{ G\=굦aňS Np.9B(NG^I7<7SD¶%bGT\/}@x^Qh{ HÓr| {?TUO_58?+{gWslv:=x-HB|ީ_[?>ssftoA{דfW4{wVN% d-N]WTy/tBKp nQ4վfqJ!KV0ڥ1lsc$#hC3`,X})BocOa-?mBB+ZUtl/%(\+Ljܸia_];+߅K'~Lef~7za>ы_tW>.w w|a`3d //W~+ i1BFSUgMPEƖl`2Bl3ߡLa* &(t>(C u*`}uL۝L9n3Q"9tu+nZ]PTVRn"O*H? t2TؽZE[08Ja!;yNzS1q\\sLH'FBy#ǒˡdA$W)x=GnlCi)FWtNWs_뼉K bzKT޺6:#E4:(G)vg)uav(őDgWNn^>|"HUnel휂4k:C` &lI Uj2Q„sp˂n1N SE_>85r4c, T v99Fo4kK@8{]ʷ\?tl Y4cQӱw.;y7F@2* ~C,ABvݶ8f+p:(x[A+@82d CB`Z3@i#VxQƀ~{!RIZ h D8pA2~{b7e3@J:6o:B-)FZ ed*VpT}`FVO)C*(8e܃ փ@#r8ۛbtpfTRbx<6*'Vn}rPLBT} c# 1FD5cHX2lrd7W*ؓh<Ɯ*uiU[UI#]aP {sJ1h!L0P'sR3HMrD˺afa{ K`#,YV*'81_JJ|"jx$As;\~7ki2 &G+yVFU*2R}I*BE|YfM_Z9NãF%ubl@x0K\% B'8(ύWJ 4@^ ދ"Y KZCNWsSXfTȈ \R"`6R$^sqk@ 1 u`8W2gҔ\HZ{5ԁ;-(<m29 4-Y`vl-ztq#MnRp-B SZ%Ά\`L) DCv "Nv5{OBwg}mv>~}UWwR .:̥ri#Bʳ,V)dё$L__t2zXAvo!;Vbۚw%U]Y-iSC(0P1Ϭ'O(UKtE|vgrIֿe Q3Qpfmbڭ,%[536t>΁uE'HEB]k:-hۯ2)v6x* M$X}P/L+E4O.bߔp<{&h=b8)q86.2t#oQ-cJw5<[cHRt )yN!=n.='X ^xo )A}R.q2b秓ct$1ttFq\{J"å-å;p{}90E}6Y)?l.5nÁgS1?K☉gt~| jWe77N0pf[o/n#$p{=ĢJV}̮3ѻLw vf`&r| ,"|3s ާ @1rw]\9Ƴ5&d(x㐮HƼڕawyIyApXy[,f1s^# _&8^$9?[ZKcFC<.k/tI뫟xBىI&:-0A:;|D1Ά}R dbSadWHU(s5'(2 <1p{CtV쑠a]^?-͟97xx8{w?|{]nMh:T~:d:ם ڲ;p{G<'۹mփ/V$!mXp[HEQ綁- 5^&q _h%P1y`1.Aau14kK :"|Oǟ>Ъ[|2s>8C^Zy; ,6';S!o~"K2#o>z?µ %NI2-e]N1h|>t2'DY"`quޞ%U_w3ۣ7NJwGU&@V *ג컧$~Wv 6x*f ͎}^﷝\cu _x -{$,:6Gc;S#ϯhВ^E3pj~ Ɯ_|0loq\v06~[Bugz__LGfi_a\/8lL}5ݿaCZvt~)񻥬_%ଖ{?3GA F|G|wi9/JX\K-\R,=:۱0L1ݒ"͵cio %~[ji q4>Ctd0xeEbk5|Inx܋y*{1^J3@*0;, 2Zi4Ō{1}sҬuVOА㧦?yȚל}dǯ?r"̓"E_'X3qID@P+d.|lXD-J ɻ\(417' C4 }ڡ̞Q+R$D 9'rse(IXAj)*Zy2:(Q]Ks9+s݀ tĜv'fsۙe˶3@IEiREi:TQD旉|$>:cfqܛ?[ H̝֭ݾ2x3?UDY8x_D3dc1rW?_ͷۃpۃpc& ILD5U] B>Rrs86CS0/P4n%{x"~ӎ:rc\s6\U`k OXGJN=Ψc_܉w"^jk;q,a.~^/i%nx5 O;vRN_¾.7* 7*=B8݅eGwaNj>"88?W_z 3j_W'~a7gȶ$04)܂sx|Ng&Wtpf.m>aRze;,wv7mFm]CΦkXT?y鏄Lc0a_qNmjrnw-֎]s62N0F{JN񘷢 +ඟ+TӪڔ3)6Ũ\5fFQiL UȜ#3`B&XJ5X8VT1K.zJsd).Ps)-M5*RtTTn0>d,eHҊL+nҚe]Us_0s:=tpc79E9l/c:DyZ\mܱ Qn!;1磃"D݄VࣿGJNI|ט].]X$b5H"VI{ ih [S!Ӂtv߆u/.nOJ':(HK)$ rv-z-8 d@W=UlEbCTc=Ay7YD^oBpwhOޛׯa[_W&aUK|Y0Lϭz;|ޛELчWۣx2}(A]/AWd`ͱOZy|֢3/}>nD[D-ǜ' {D(Nz:1z)91 d$_8yzM?H+G|An [7δݮlxVA/c3zL|KVw !\i9zyP/ž&Ja=Rm,D*AĩXZzÅsnb4;40 ?νD:κu4dN#JA`Ε!+'p'*`"{*PaLL-,>tGXrtsY{ĭ̏|GlUBoSg2%w=@& /JaX[•V AM{bC<<{IƉ9EsH;Cˈv=;G=?0q&.ro~Y7K$`1qApz[AQj&| X~贳ҝ{s7@x=o~ٟP_~#bnn~y}G"vak\o[']M>7S~}կ]t- ^1bWK|5:T-ISQk7'27- ﳻ>\nw$k׮ij^.Fݦ#aط=]?y#=yL<_?4{}KzJ{[|2tXY~g}\0GjrV/4;X^!>V֕8=v MXb#]Pl*OJvTS`y!gs[!-ש(V7R#z*lF n{A{BS(ц{ XM)uUB*Yc9_R غdPH\$}][I}է|\f$6d73ITlɪ Gl62PkjiCkcTMlqKkbZ:&tK vȎ#߅LnќsĪo"&^tϮ-fΕ1g|3 Df=RrbwZJw{ >vWCzX.*#RI$5Q g?_-RKyd;fۨQ)`2t4"R>ȒHg'7\e.gI۸R!H $;71%E )t丂:Ov P81xz&moP{( .tNݯ[N1f'&f{ذsT5K4|4`9&pɯ z=iUdq9r| O *CRBNkq1َΚZc pwGDK|D4fz^EӇ;7*.K%s#76:y2rDN^ʸM$»L1c8n"v9!pI朁!?",\4+Ux|xfs6ã;S|2g?)9[ tyH_&`gm4hߙds6DF< bC葐sZo{s/OҩOS]3X6P#ϰ+66B 9_"JynԼM.դC d*|+X‹mR4 E_>&lnZ^[94Bⳣ5ND c(R^R~T㝾D.? X#h nQ/\\'[ޟѺg:V,YH]9#CN~nHɉ9=E&|bVJ.h#+Ѧw#'LI /Ucc|jԪ:J޻tn؆=F 2i2[lj(5 RftGcF+*3ҺkUl![G9E{GT]W}~8v&JU(RYsR"uR*9 g1vR58ݓ%MmubciNv[T9[6)AeFRg˩Є:EFoC\ɮkMk͊9t-=mt<n> N+q[:r٤s Ep(z$fC=F,:Xܷf.YQ{XL9xڰ'4Gٝ4  1&wݘl~Xsz 6nTl+G6,:1.ŶK1pĢřMV#bCѵw;VuWPl_]2(:jeS\EUG$5^K饵bV 8k*19wun8_55k-wXǫ )P<1R)YS1NRDXk`=hܡ2s*E*6f}=3E[-i-V^ƪvTmr^)գ|7Wu'ɍgo9}094Qf]tdx0tt716 :v#1cNPЂBо*;ZF ?>~b먄KVrKpGFz lRN=R&7b8tV@Nb 7cNmL S;v9,'~~^a?^z7ؗ~{Ȏ=a]RRC<[5v%x\Q&{.=Mk<ܘg4~\:-fevO t9,\fkfNxfA [ݸ}s]C]ݳvnsyH`Fl0iIJ@m5}uP1շ_Fz#aPT ?*w6RN#/% S}71z[5)9DbP8bRLA/M=hjo'TSj$0Nh׃D`526ƁMX*MA(A?kuu|b)ġJe6wLNᦢKKG٥P[DsbF ^T5pFj{>2 9&Hc{LR<o?$>ͮɌ>2n"˶s9]L` Y 겏x>&1vۤ͡;zxk6g WJUC@ؕrQ-kNAc:ۢXkKjDE 4Xtq++ŖL荡DJYx孼A,r0Ǡa8ǥ&0鲊SlP,䝕IȖR?\6g ͛>JŚlX ) ,I~]裸U] XXfͽExT]ney^ܓC|,/>.Y /t#i FUwNV7FXmN#>-F Zk# 󜫗gqGmz{Hбe.y{1'-ekED<~=,@α~-?U{ԇ=Vh$1'A5. M6@?Aj[+IY`n ;>a9_$G}?@3 .+Lfu5ڲ&y.{<>=+tAR]UM `A2i_㭢, =p= d]s;K ]9~SX;R"J/~\\Ry. kĪ3h%N׭c}T AuWvgF8(RXw}!%K1Y}gB'ۋ9!Dqûv5Q'.zJ{l{Ob@c+cxcu8%+2Kj!5W_ˀXE@u_ӟklV]FC}_<)/bhF6gQ*ZSCTjz*|qtQcE\d=hؼtC2^͐n9R;%feTMp=z=7מhVh4cYa)]?;@*5c`wC^'3u%ur'=+ š\qcJƪgw٫߫L-@\{ペ[h]v*Je>W c'ɡ]Y)hLP-)^ؓVݎV;g|f2F3eEMoHiyi5IiKȞSSLP ]t:4SvDfƨ15sA Vr;SZd$juzWLp'-4Hg؝SA!ڻvގ]!OV#!EP]Б/·A΃#:^,{־.g RTh5LuUTʓt{gOhL֜ ŨH#P;cl,@NQ:ܚl͘IS~3iFj\~O]Jrm>tbuUUcl{`'tG6"gөrSX u!9ďOFGt{׶_©7˖ݫ!pKoDc ͒Yms;9 1Uϔ.]q ~;}/uA.NM>-D2nZ+7!QO^l] J.$ejk('$K hhظ%n*&`SATQVA?Fɖz~YUXK{>fcR ri4b #~?[v灍$Ȟ. * dT$v*{Ezl"SL0n5آAHq|r/+mB>>X=I4cW}EX.uscu0il6~kr*Lx̜JD;d>kQ;`/`Dc򸑹_[#/2wC i tAb W}ȍOr>i/}ڷS!s!u_V-'~WvhTF4q&eypwFZNo^SJ:և0§뗔ڒRa5#Ԙ \60][P``Xӗƨ}#-DܵY9̻6~y9s^$XyNz 0 Snb[ͿVp.V~=Wrr5dΥؽyחa'_ Ak=r,o5_\ճV35+ו6ݯqrg-wÊrshͮ_uv}`c!j%~{{Gfij=Ogo߂-py_ ?gàuƒLz$]ƀIlP&R/ɤWLliI/nYD u:8 [U.OD_jJ]}Xκ5^GQZ5w*x=C^N!^v ۦM)7_>MU7j6s8!eC`WlֶQ}!5fs>'j53I)\4/W%q nC8.%q6j %b,<́Y@uf4,mRg:Y8g)DHEEBz"s2ka&kb#!QcP$ֱg `Hݱ8`EXt/\U@%K1Ir_o8tOX. 17U` .(t>B1{=ke'l/G(Ǯl-ĚxDPrI`q naz "0g)y8`pQr K\*_HTؤyd+O2ccavaIƊ (@+qRD 8Ef]5i#;qf Jc ԔUbBSϐ;0bnԎׄLHaV8Ҩ"h+QB;'jq";Jj.*ω)D$&*S:^.ao6ʩ*ڗ;e?Ly3~υaB X"l}*/ZlFZ`j0´o$o-ւ1~c=YɽλW~P,D(pIlG?4VQ2;Nnכ|t9B `pؓCuI#:j;sIq7юR߅Y`jz㡘 fcsD/=>Kz/ꏇ1FD:K(Qԛ9DQ +EN-QQe':&U,#Qʭ,eSa+-I, J-7~zDɦkB`oǡATu< wz,{-.e+doo K9AG1:"yWNϴgAku?X<{Cry iӵQv :x޿SZZ\r=^ZEa^0Au:fL8cIMu+Rk`Ji2[EWf0y^*gӼ&Ԉұ7;*ɻa~펇!-a4?èp|&WzpO׿> FyD(˞ϦKg== lOKeQ6h<1˫~7zoL?O5au﹙h""4m,JK(8X)\j.R_~DIUH+F&1 (rBLP8Z)Qj7}Q|/߾7SK'v"9{`#*oFϪ9:2,ݞfezթsZ= YbTө |+[l(J۳n%醭$ݰt[Iʞ ED8;<&ata@R7ʜH=\,O54}Q#Ϧ˷5⧓wWOeMdi@/gd.faEI{4ݯ2iԞZ4/8W0dU:j"@8aCا"X,E9)g~{dQ=mfd}nAU綍R$gNRƥ)f  Ywsi}FW c!E}(lAsVjn]zJcm~kUSǑpii<f|'N hɌ 6JNA fPj3S9bY`rRJ2) uFԚ'{'ݺXMBFO3JP*0e2?,x{fS ͉v *nFZJݔYgsH *7O;43['\Y^fzW&mLh%;:њHBb07`  $R@,MaHS5)Nab݀bTd~粐6TGw@65!)U8%. yycc&w1l(j m8d4oH6"+*Cj^qvGuHu(Tы/[޶Q{?lFcԨuwifT&0-pzN 9"-Ȓ=y:zomMw3EXh ("9'X!a .Ç5B^`z_Ц NAvN}E6Ǩ&K(oἄ') !FM,II ݰdCc"xl65!04]lF RiMggۣve8͓"$ŔH&׸tM p/> ``l1]w8fತɊgR.$ rZ:hcTTThœ551וoכfԲwﭭRghӬ:(%ԻV#u) T>ʏIש|]@6oq񵹭gֻQ˜F;o><Βۡ@/zm׷7D5wz.֫|X ?Mlb9)Wij] J V_}pKV7!kek},n=Q4U72[qnm;~آ IFy[Je8$/o>y8ᯱK\,aՔiq 7B s%EPxk}4] e>&d&k. ̒K}D(]{@Izbh~g#S:>4Cӭ(mn7%fxs3NAjVl Iԡ렀b.u^MvrF 1',:%(S2i2OSFO܈htodL;hߏ(*I`(T7.4~ 6v3MhX Sah[eJ"nUAO^)C|qRh,q@Rٸ>#.t>xEҿqUϷUD' 4Qiʳ k敵ˬr$UkJu^RFa̴IS~ԕkH"JI0,)'@-mSf@5+3بʱ&)x,@N#2]sT*cޓ$W ط5HAahv7me#qAVSEKVS$%9 HT:#*SAsq _>0%<(i]bwV=kU { TtJq NMs-yQ g61Q+Hl#"9oVoQdZXR?4@qZ3!DJhrYUӄ'A4LbʰE)`[AN3::2Odʴ(j弰<5R(bZ?.]@,'t]9M3ZG8/Z*RtqT}8('ٛ7Eb)uh՗jG58UP2˃z2P~Mif ,CMi [1Q`JDaUJ1D$6#sf)͐nJSE}kT7Zfhʵ*֑  ٞg~Q!n#3Lh.C=)wg"ڎYrD0"z#tm?+"Κw)WcXiWTO ԺƿIƶRJv΃ӳ=ǹ,IS@")<#OpIMX&Ty!)(3gv%!"=(υ8o;?aoFh;9~61i奜 $4RtJ KEc'ET FGqhZ JBtmrV6oF:(=G>jZ7%`9,V[FgLtr9OQ2NW b9# f0X" iY=t*$/tR"Ti6"3q ũV(a, 6bR+T`BQ5>n-GG`1(`(bĻoRb7¿j"S*L"ΔS4YNi!a)}2+a_ eq_)C̸LL ALfj.JƒcIkhTg E}Q/ߞ%D%]OԊ L9m(*Ů=mճʖ5ᮀKͰSL@m2}$E4˴XR[JCɼpek Rx,mJVebj,9=qT8nsJXwR;JQLjj(Xq0Ŷ:FaGi2жHƌ vI`}( *d $h"uYZ k=ӡJ\9̈́bĸu\D-ރ򩼴+R8L^RlPL5$sƊpD=mRvZ(m䡺CXVH5*N*5sNΔRDoYH|l&LeՑ[ĸ$5_jyڧ>%7iFP܊ E%t&ȰVՏ.p0'tp5#T8rSyYS)Hi2g{d2*52yF歭\ \{ ?JLTRGRwc#^D}'^հÜN`#Ch7O+9@3!QxfBtVM]6cx1ʱJaF Ĝr$6R*(A8s!dg" ,S)N"H:@ZTf;jHC$8TJm+zY1)΂tұz7୉ޭvhuJm,ѯ%v~ff_7ơ֟>؍DYjY^P2 `)X ϙ^gkGyl335&UĂlv'R*N`)9N^Wzɟ[iQkpQP,u0I] aIj@V%6\RO3RfQEWSGR9֚ԊP{j4Bq)ߚqT[hXީ&3ܚ%Ob79j[ 3f(s-1FjǶF*cpbV>E:\K)Μ_`*م"\gI@FZTtvK`K$C_Η&BKe&JPʲO#-& LV+l2j]sWsWBcSj twH;`b0P{WְR'W3OMSa*@PHAea Qm*s{*-Oʜ|s7mGԙu=n8 U6D88&U~y#1cn-f ^ XXiq9fSꠗqxIR͔<9܄ɨ[s' f=d@%]&Ke_%%.pn 5x,h6eBY.M5Msy9k,= kq(W؁$жV,FLOܹTD ^'}~J!nZhX-"86g'?a'jV(Lz6w՜j9-|V Zk)4hmOlwon$܂^#]v;i~ R$ֵ7ֶyn".lV;[=|O5]Cre vuFVF.2.60\*9QQ_ NûsPOrdx`Ⱦ:=dnz>}yOz>Ė,j-CBjˆu0a'gA[eŨb^אR<7id4'ևF` NL72HB_j'$KOJ:nW9??/N'BY"dwfx|<Țnn^;|uʨ₋F ֑(,QT'pMRR1LcH-N8-3TI%!xNy.m\5COFOݡ=uooyMy{{^:_M6| _*~I7Yd}}Lo1!bA\JEۛ#Q?7 YdSΙg<ѢՑO7ŬIprXNSMut^_- hbf(<yqltM쇷~at>^7rރs>o 6w]sFWXp{ y?T^?RSqSJ5L(Q&);v* 4 ZK3ӯu; Y/`g_a?LXtY׷ ~6(x drU|웜w ggbSkF)+w}n@?&0J`v|upNg77&WI};E8 `?4S\A+8VRciz|9}sg:_gXM^,1G Wbo;# 6 4dzQn2nFƺFP)I<4ML;jÁ{[H٠SyX aKEgC PM;/|X_v0}xs9<' >ӯK}~R7g! O OHFEFRLqLY0柠o ?&c7E9wvGw#!łw=qzqA&rkp& )0h' QTjp}C {\_% Os&ӧJh%T28gtg/Tn@Cx"+dY_=SjCx(gfS/vizeחF8SWL<ǽ>fLc/(Ӳt318b9=}p㘂Y}pckpc͚OKAe["`щ&̇Bp3'٣EEG{٣EֈENG#=!)$ "$Ke*9SqAH"1\r1FXj05`J۸Tqg=OrFvNvVfaR3q:V4eɇY*FvX g>+C.e7]T0tmYRYe~ǒǠ#T~Q*Nq2W o\YG b+4ѽŸuuňhRz] $ iѭv=p=bL[e; &@ 2*Q c#GLVYk7/xÕ;[9ʌnQfm(scBptd,\6ƻΙܒ͛.neUT{&I]l=LYG̴;T}Ra9N*v 9fr>O*s>d`K!:dD Z!3oh}o?D8@U$$0LxWYݹR\U`ꎣG99Zb2Qxu+dzbefG]/~3) r%mCX[ h/{gH10N7p@ov-lmC;mȇc#SdIޅtoU {WfFGIj2w2 &1r+X_4>n zgh3 /0%ů@ t!_Z_Eh-~Qn-> ;q&RsyoRJk)B DNf`tKMQ uVԔ m  Yq=\G,s0{JY0GLY{kƕh Bz5v)]Ϋ'P#rN8@^cu"kDV{C+#BT]0b, 5- R @00CDiBKjFRckl8RCD,ˊx/5=EDjJ5Yȣ;|,E2_-sYgVN, aU"D6Yy )C@X":tF:tśpoNWw?\7s/@y;ɷ'ABa"x yafhDFXY24`C&`VTTz  (z&4f!^o3ih5Lee ԅ)Ji?âך1ˉn [rn ׫l40<Nd|Ice.CN5`?59ꡡFw _+x&}ޛĥ&hb B$O ;F\PAJIǴpJrSTxi Q8sO@p?R<kjEM51\lണV[b?CLL hHIJ)FʬϋS IS'&`1o`@g7 >R6č1m=X"{i1!U$!N8M4I9N#sk@g@oR2cE&-HER8-q<8qX}Ixш),3 г)Q|fUց+q,NC"(0ֈX *-2] 3šj  hJ38!"2*qhCB) t}ޥQ\gEr^Do%(ƻk9|ukmgn}chTF_lhF:b~Vng>qv{ף5ĕ1Ph\#$̠뼝K61Bl֬b.@sـﵒO="GJU)ihsp0v b9rݳ!C^cb68k49Mr #RК`aZ,JgBІ4K`j<Ż5(|Yn}5i8pO85#Q9pghtԘTEb0ѪK/5DiE 8f 2MT8(k攵l,.HD cb)Zn;A%`=1G2Oj?Rj,Ug35"CZݘ Z!T \iъbdf~j~EAޤ! "| -z;o_))ϔBz1%ynIyD՞4 }n`x "ɂ( X ~ d:;š7`npWLg:?^_2/WOͅAIu~ _<@SZ\,=Ϊg `DKik}Z؞mن%ɗHՇ/<ސ>稏7 WTVEa%IBQd)31jTڐb$/_)W7fTp~N1i^Wdx&}wb #We7nyKW_ostnZm TS-לbZ87|-AJJ1u.n}$@5ِ;&oEyGMrYL(]3?B*L;,_j㒠wT%U{mQ[>h1s.V/c_FpB;5Z֪5_k4a(]ܚ5 Hu E[%D [ݾB(],@~rj^˳߭ۚP>k`\N_ۭG`cbXchr/+[c ſ;WQpˌH=9RZԠp{Z&SF d"!:cb h!l5E@VBEʌ)d5C*+L-˵UJ2(|Wf^OWy͂A1EfR={<{_v|;ʊ77AH_+|s,o gc7?4<Qn/sSh]zϥXY\ٓ|"@:/O'2R*"$䤎 8Wa)1d$?kKpBX s9X3Ne"s-ƖTh*ӨAnXNR!ga`FwP0 x=YS%%YjKPUĀKF#e!FΏ:1 P<"m3YHn z(ЙmRթhLH[ȷ5O՗?NטJ叟"t,kȟH=JSp0 1"{BV"oN**AtQ7 r&4 sH I"] ($iYz9xr.9| .mĵp1wOO|0'ad>Qgny%7<!l+b?x)2 ZkOuPTQE0h3 7sWOV*VUJ(*y*rWjYd?f:^^_ z{!?T嬂OK6?m)v=hW r #g8~<o§Uw|27Ev$b['O ZMzA`sbͩC6O0a[ejjl ;yBGv.U< ׌gMYs1^67O |6+`ԧ %*`ٝdVM%^Z;'ܱ%bʎ-ƒ':wiƊ-_UXs`׮t|Muǽ+bebKߌZtolFb:+;/\RuW{ujOxhKMYʩ4e&˴4,N,\HEcW@cN* YšMĞ'X婑Q)qiИaZrE&&ի2犵ՖdҀGR1q%ZKa%rYOPdhXBΘV+c߁q] Y;\'UgmU!hs? }a)>SRy1+#4EdY. $5!A2$IEqw͗&8C݇B7ز6Tό$ GjNʆaYzX]]9ACr`mֆs={>`]{RԳv%=+!~=J=D #ʜxȡFq$؅h j 0@2ek[bܽ,a78|vtd퐋hSy%x嚹3׊dsl¬o*(PAEѤ|SNxGނ'Gn6?X zݺC%9b Ʋ.D>,nna*h[ ޠNaɠU˜k̥ňé0F3( [p)dh)430Ĭ31NyG?QFqPRuI@@ Ζ e<]NKt\0 wO5wal7: ]W {*9nv)l RG(mė7Q:BnZMk[7e9Fȝ 2zȌ7Iy?29iﭔ;o GlY4\1\-6G4&nemX|&}|kN{].]VxV30L/˗8M=8e g/7zdEéFLϟ?r$IWg9bTFpk\rG/~}wq_lusn mYM1^_=FQU8੽[^]n_70y sbZ9,09G/ZW?b6QàIY BzdL!åi.jFJ)(o: MߟR0mejVF 4LxvIY.IV&/`sڥH^0 BĖ`S2蜳FdjkoFk`4z7RrP+L;}ޅ (못)v!Y2WŴ!pzƁYھ[jAs 9V^+q즺EqMu W҅MT)SkD1fkqRղRᕫ50ӓ e?)L\;4^9{Nsv`=1sCqUFanU=ꅮ;`gCU2bܡ qlp^fpb6i>ld4@QOFYYֽh-s5"6աQ] @ mŤ3jX;M%,!W&3π¿XL#k䫻j>5pg !ཀKG&#t/JT1f;iGsRւZ DNˁ݈.)lۅZm7#@λ> |,dٖV >ɧor9yܾL ʗ?Dg};7 ҂ep ?*6'[oq`#5VZd;0ٲ8Uhթߝm_h^|9QO_Ԇ!zEmfa9=dnL2.>+[J[۞{*^aB˃EflJZ9Z-k_CuF4k73Z%xK7Yg$iMSXk+hZlCBkDO!]+,Mǻ@a~ ,P\zq*;sZ|+57>,yoR.\&L7?v~zw)og8!E^ Lb 5~SF[= NѮ(Ĕs]≂2XBk-S@Ƚ,),)P$$eP^- ╜ \ .OǔpegR|t=@] <%l}][69f9!pCzw*fp|ywk`pxge KOyBDOi2hD=EιO(и}U-w= L^yy^Q#9Eٜ#WROD>xqQx̞5||ƠGH`|= cҞ#X 7Bh m z``;E[zBEdy.r!TYs+-Ȳ܈2lN-u4vÊ۞{,wtFQNqOFXu81$t>aO^>gH1EIRJM3ɚ />0.lQ,n˭l#TE{mM/7ևV/_j0}l4=FրKz d!n[ Zk[9 ֥ NKK:)RR>DK6 IGW.FαMF[tF[Xޒ;;@x'qcZV{}y vU:M0̗ PpU0gL-ؕz1d2m-6 Ӧ[_ μߟ-Ewqa;h?Pe97۫DۛϷ7S90wrDA~spCCLM/w_x?+ 9Kb .0#;@U1|s$'4BQȏj4h*T,®J6Fktk{I%I}~uu1ޛEQ'}}*z!)mXwE촠w"fMfƣE7 ԏ J2߱kwMyzu7c*hE OẺ\MOTMzMpcpSVuxĦ&=8eȍ5;#7 or>"z#lG.:My9PIt7dQS]2xF 衻4{qxc:[=U4path kmV<_`=Xf t&mx,G'/e[ԲKw!A2cU_X*~Ҷ}GXDsv 3BrM41TIޗh^ b{A$o7U|@J@$fP4+S9<ɬxމ]"<f7P)%DRi lJ`E%bQ_ 9s{EY7e,Aǔ8 'U׎ V,⏷_z4 RwLZN*6DGb]`Uj:ҳ-I;tr=2A</oN}k? ű;8r0 ׫NTN 8.X28B |\]a|RKUUo&a|V;[n_QG79@H=*}ͪS{Li!ͬz)T"jogx \T -=ְ!RZńR%2#8+^+Q! Q0AgϢd^b|d&Q*ZH+bIl^[RUm8wa)1HUF21@t Ѻ2ThV[!u FVi[S2V8XT.UiII xGƌ 65^ F@:J$ 8gYj4o5<[s99D/VYo x%Xo񳊯tpQTk D EJP 5.(DskTMP%{"}P\IֆD!1ehbt7qc'8ų}h'T'[8kmis➰:<772,d@O,8к 1Da!Ӗsg<AȀ?cȒ&"1wR53sI/ v.ABdX,,fǫ7Sl ~xN=nZY/&M3{Sw/u}x?I^ɻ VhozRzф >~!:F* +b #Yn_?gpQD[% AЦ:^ L0E,_@̅4r ),I%LKHثWc>erap9c{ ;H:.A ?pIL'\3MB %هZ.E]u ;vA.1FJ(%HZD1}5"!-S3] ! = C$3W%~.q`ʹpXٿp\4--}@)d"VkIEY(% TP ƭ`e.~U5^Zl^Y&IMk IRJCD  Q6[[WB9.jr3^"ŖgRm$s!/7MIIU[jxD2B8 4*b<)UfM# dN#  [LY-6{ͻ&|۹gr\e.%nlk;;w'dX|0_dX>85Mr^d@2&KT"scl2pI-BMJo~1&A'H>*@ ÀJHkpw2p2A -vSko{yp:ș@}Nj&b5? P5_JkCE~խϙ+by=n Hb`sfrJ+N,|^~m?l;VK%t[ CKmkPkSs[C 휵&nBլr[V[9?]?i(dh<_D}3|*Y1d=c?R3P Yf{H|eV=~hv/ygZ1)MYR=_{UKGZ/n:aXOw7PKsP #)7 q:FPVQbmVZ`-`UX[U\Բ"TbS!feE&>Ɛ TO=asf#{ub4߭jjCιz&ӹʅuQ7X,ŅNgfvJ}u^hh~&?݄MA;6#oz<^U#az`gj*&_GS͆][7,vXkll[ wefܺ.z6tfAɍ$FXx7Db[BL'}[AL-zk[-M4ʦ#C9wAľwsż[-MtϦ(qOOoy‘=0.|9Rc`!\^A6ZYJX&f"pfk,0ˁœ!34g j-$%s4%RԒpa1 H>*G*WO;㞩۬Eה"mկ&@>sBH-L`egN5s~vUX4R+ѐj]fSk>O9(f]g܈jW!ɉSL6ڏ4j8  0Kkn; %L&Ū.&NcO=_sKWKaOqv'vlر ŝhUP 7r ʧJ{0蕑C.gH@<[*8ϫ j`qORͧ8#3lu6m!AWr< XgNh̜}z)c(vTJ] y#J#Tzp`އ0>it~-[_5e (wOɥLZHfv1.$Y9xפQ@_YTub#4*rܛjyɾۆxc+t<߼zXE!Bs4oHUU^01 ˂Ӛ I??\ݬNn~|D{xwgigRZ?^^r?~s.HO@f ;1Jl;ꅠq[߷t(jۄw_t-hPBdd(L)g)Ц:-t>ykVOMcu hb`veJN oB~E6R|MPHyՃ}5wN6vSZ% uqxLn4.Ei p{55e\!n.H$ݶ`&)#YmUznt}Oia(wx,Ʊӕ< sE1m4zj%ZGhR0g$H'{D;( rJQgPREmTuS9DiViXp%fGM7~ۮLX&Ƅ+ؼ'a&n{ʸxN4+zIpȅ6tk>~'j~M1$B$mk^+ iAHpO#96$[PeaP ")ai11,w0X W$@xǞ))tydMaltHRF^1 I)sX~QtzBnҌW-~M)Br]d4 `6eA)cSnr{|j7go>Wv|C\G#1ِzEH $v4ߩABP.H;e=A3T (O#d4{EcD@@aߍ Yj{;N( ,W[e,` ]$:1pǒ`3xQj:hp@sGHɑl?|79Ñ (G6%hGrݤsEQl5f9_rqnf9%VVXn[^q=nSXW;S&u yEAf=Il2t] <ĉ9}AC}^ s<Ӝ=aꖻ'I41Gylp;@Cdñ|%3̱%/5NKN)5XZ2п$tGE9.ٿPBT_2NvL%?DA9GzIFOrإ%=w\\28*rB˶<)hM_~.qHB$JS~@gZ6BAoKyŁ&HK cu#K:zMw-SC4ko^ܙ!5{k4*֏c jТiԺw㩵hnw~`E@wd \3jR<]w.-mtEMDMa׍Zw,f+v>ڰaJBe>3_6f.T-+B}wf0~- r ?BǚKe:P gÁOb p%CB > m Apo|ƱHT籷j+ʫ'ڥ "XQV[N. qpK]%2 ,rt"X|1 +3l`l26pߐ]Ru14m\7P*c4PbdlR$Ê{G0H1b7"EL~IR688$l! %ЪY{z?N2tBہ@P&q96D#6tdJ2Td&˻#ȃπ~׏$Xnͫz;pCh0DDH2O=Bfu^Ugw l` |0-ёT#4?};zP'"1.G>xfV cx&I-Gj BQa&}GHaE=1AZ#F$j)"LbCJY ^/$CZۭN#5R"PiS"I5Ke8d,UTkIi!e{Rt4`9Y5leYʶv1߅X0ǀ d"\1BfdLy8JD,g sV1SID%Fg=5 ͬцOix0V;*IQ)YIP8j4.Ty\2/p3Z2'`]li#MI!30;&=:UFsw|{ToJL6͍iF7Jygrٜ4]&س_]"waq{N-%Ļ{oxƱ8xL19KمǟmДb%S 6݇? Jxgr9x(ogZ ]Vq7WG_[8pB'dKΖEZoѿFͱC7`wQg3\"HZq/h[7 N!Ax@ȸ_Km-*u^U= Hw{gVENk%Ϥ{8W_=x.n<+ߘ=|kE&<["]6JF?ڱI(OGǍ2A]|{m1=U4HnOi=,vO w=:5`7 XP{}w(|J͈rP{j4ݭhQ)xn]"@~ 7Q?sZ=(s{X9m$h$|8|%SVʒ9ā͒h8CS\b<ǝ;pjJ8s@pT_DZ|ג>J1s;ՠQWb~L5 4U$9(f&:5`=K ^`ܤYI$ԦS%Y1uf?y`8??Y׭jUf ~AO!M _CTk]jMbW>^E*4.<R4~-^XK^a$ DTKE} %ScXBd|AIޢ$UdLc;YU84eZ㊙!g;ejM.Yf|/p_'s h*`!YݎC2{gVͬ~H>:F>@exYl~>`JZn]P|C.SW`ndףeyFuVsI,y<& EL]d*g'Yx1gļN"ҙ @DEC}(Vz7&'W3:! 4`ZtR}Z2^CBr=Z`.Y3|~ODʕvtz٢D֕v-'c=Yk{܃܄O _ 8aʷ!#Eo޿2nGSR uSoi hTH 8ED O?qiSFY鎖F'5r#.'&ZWxIU Uk=@vnИ 5ߢp"aS,Bl}ҏf/2I8 w{yNײK]vTɯ0qRlT'g9uY%4U=2eJ koʫ˫iU2ھUf-0߬4jPEtVׂ$*Lơ2MBCXWF;ktݓ?]~z7'fX(Ed$Fs)DYDg?y~pYDi`;l`] S"!=JsP]p;{5t9=HEcS\JlL3i*SbeY@(;k}M9eEL)ue&}/zmiJ rr۷}Mv;]>"k@tH1a,{T0|֋R{1BR"շ+KwvJݏGslVa+wo'F+ aՠR6$,ș@Э#RHGaDZZ_umuYߞR"KQ,'v^-h.NKgJ!j5xa[-aњ6pK n+|AuAj . \C-!) #M_Dbjy 2KL-O _[.5hP zM(}D>@ ..01G3B˱ϴ"5_.msėyBbs9VRN 1lZլ(YXߢ*{ ~&r_7G@t̫eHφ~UQ2{/ 6xnh;cIMʔ1B aL jWbX)f nR n?!>~ ԼD_F7QVfhITuhY BNzcYh~7 \7uGRjMNYQwK+ꖉXvT}qXE0b0OH*LZG7RK;fMA[EuB~<*xe䌬\Eϫ^feѳ^ q)D4kI–7]22C]^&,"I IpI)Y̑MI%QW$)<|:nxʼ#{dy36!$VkV͚X2j&jV')T3dk*ubrG|jB~^+-TzdywFJ# I~*;rߐW{f#OXG1"q!ԷDAR,Czpo;C3P;%G.F#?LpE 3Z};zP'J*cg ? iSΜ!:l@u%uK c{u3PM4;;+rάظLFe27urBxF:t` k;iӋ_`>0fk)rL &_yJoM-k5m^6 2\G0`G vR2+ϨC"R`yڬYYˋ ʸ$baBp- ITKj(a9p B("4^Nߛ+̫p[7JPR'{YO5X#Z6u\Q|jSXX+2XΧL2u*MG=OMPPxXqQn׈ l`U@KRbSA\lSncFFmLJg7q\ʐVFI`,UkjP@8IN<5@r68׊bvQWbEl&pYȺ!}UHϗw1}y :|2u!>Ds?.[4\/F&Gl k2r3vQLjcͲg3v?GS.ff@NQM+=VKG|o \" H8 [>|{*q2vU8(M3^V vU-% 2G L M;@ji/ K AcbȒOaSh=k8ffQVfv<ζ$Tk.$#l_w(m YK)58USЎSe8q_ǩqw€SJwr*Zn}p! CE'a2m J}tuċS Dp:L?ȏfJQ,m#Inon-pL&`2 f4&z8߯-ZT1b2IUsWLilج!k=*{A:oJlڣO~)LfO~فeIڕS3^e@##LO{ XI,H S _'+wMr]_~|^1)bs:]ݜxRD%3Q-eN&[r<Qo7wӏy r7Wޖp'C!KJ~Aҕ!+#zG᤼gg=t39<5=G?=Uu=ti7qJ;uOlCeU>CBC8_v9f\p@Ewzhkx\zV h2tn(tF3 yGyCGUzR)w3.K\-{IP6b˭:ZZoSw#^TF:] !Z70 ʫlOyEJ4VDEl]g.ei,͏h MA$'\玲6no4_U .r*Y4R'I7+Z-uX'v&xlJw-Ѻҭ Y4H/bMUo]]C~^Q 7[=L/- vM/_8J,%Y;m]HgWreTHlH/ِcd%~_r?{9c?9=rÐ-Q>,nיCl]ϼ~ݮGBogFF&}of :AD-/\?U/B6^my ϪOaV,^&HY6 yy^BSK(_!3Vkopox{GhD; j UZvQW:g?QkD6j:1WzW赠 0Co^2YZITy[b@TCLDH"2PfzIJr}xX,-O]$D"UjmEe!j >ZJSJ:>dலi},̹+D>Q/ _[^i66Rëd7>/r;Y ' ;3w^)`6C !VV~%R)~R OI1<)ST ;A Zw_ٵ6t^ySv\6_qT h,% ML )R/RH)B?oC>a YmҌ.r:`Q=zCSr} 0~ ;њd _b ǫxu]m$&Z5or9\K)ް/B`W-MKv HT{XL%o&L!JFg~k2pՊR)"ZQXBRRظ)\%n3n#I!eҖDdhKTʙTT304JRa7O9L1N xN#eb{fNbǾߺfFSK Evk~p)3 4݄c (JIq`:k}H8S(#b:SI7jS`caM)Rq]pz6SqhF-& vjk@ 5Zss*#_k:Ea˴KR QDy!AQ"7+ |ahj%5LHJ<)&FcdJ%p ISF'hIJGI>`a5OHD<&5DZ`IT)UJ'!VB"6)D#D"8 růiy+~,$DžK/gH=, "rKKoF˛¯COkBO B!Av)AQ۟8|XM|؀|PŽCD@SS6+@jV⭙7yo:z P>q*pjRP[Z$Ez!9kDуǨUG/LYEwg*p$Lt^/n,;s}ƌ'.e} N *dM:IeOހn@Q&H &W˸.|T^bE&p+fJӑ}Eڋpy0^͋’"*x&8-$K=v8(jjp,JbW#n~DM{^g`$uxF/txڊ|Z>!pM?[TԊX Nӑ dˢ` Soo߮a@Xt X]_ 0PTUJ]Z_%l50Mp'd7fC[(C[H^S{b9 3'y9]VyË*3PP"ɥr# Hᆭ!|[N@0s>5 *_1lSRm|m\cݠ1M Oϲ[ Aÿ4êY:QT JUÕm-8DCb /M&&hh(!@q/+׀a]n[ Z ">u吪 %"6V,1-U^My&*\|8LJJxyآoaEYgIFY7P7 &4r4$.a{w2^,t'.ڢ|1C4,{ GQFpqZUҏH*3B~ 1Jտ>N1 /)O7ACDo7ҪHq*WP뚾e)a<[?)I'pRm<-iJucU܁Xl>->5u):nZt%h4b k< Z+C"$ݩ&KЁ"ם0R=_xQ@#jލ~ۖ;UXd( FR|Fhu30[|rbbRQE- ΖkU1~QCG~%BEcF%fڷ?nݸn._pfwݚ9ki.5룻p]nL3wK v1c!LRhTw#|7-Wb\(m {Ix9ⲋմJ>S.-ZٴAН&x1n"c6VEZ$w} ׶j r]xE8Ne?(7Ϧm& Um`WpDr엫~q[I:2[d+*#,?2ƺ﬛]fAuh:eTfF`C' fԇc*27n2D'Ù( 'L*2.lF2~!uGL]&b')KO:zU;c:`FV2ٻͲ_D@)=FQn|{r+`9pݕz\~Kv# $Hv_re|qE Nؑъ ubS4ĩ]S 8NmmIKq*4=pr؞ D)PC?k!IAn/ɧ+ 5):A.aqQ2p[t9gx !U™nN1TGGNe=$mcHfBuۓV\D2l7hG@0P&Zq`GnPH8H)Nա 5_(][F++s0YHKG K-UZYbEwm/&FN^Nsyi]ј UDo|%*Qڧ҄-p.6Uj+5?نù} Vڂ ܞiRHC/>M|e>k,$ {l -úosdࣿG[+ (TNXBudsVZZƒfVE֨B.PNFqUVDrAG_•u>ew.aQ:2kSdޑ@WPGVڅpOV3%=XRa N?贈bHJן Lg׶% i*b(iE(DZK_{czw暡pu."DJ~e,5fݙT@2<*:6J" "#aN0{;yɃd1m]qK[02R1N5fDr)3/cCmM;w4CԄڿzf*_$W-B3,UH8Q-FLGSۛn2)EǮZQ9t$^QDAeڝԛp1ݶwo+ik-b>!_@ zeOU3;pyGZm!Y'e2W{ XfA*FB{cKa2&D>2C  | $1p.MI&p#WhA&DhM '^@*lƴ|>Hdm*6ڞ^##>u~A !}8$>Ҵ]*)zw܌i2W ^ JiKfñ=<.DA\1(ۆ*K|'\aQIm p#e 1 z[8Љc;7*C*uݻ΃5:k1A A6D=jxSE ch4dՆȝ "(qm݂` J@$u9t'/`6t+C!pJ4jN=ڻ98QaKL α U}{j6ݙRt-|jLmO3AC7]OJ%vpi D@G] L2C.?֫o?s5bE8*I'#/l7`RTТK-PVJȖCږ~7`EPmIsӟ%Z@R{ o")E/94l6L$ /V\b%۞;/ 4yj_56˧ܟ>}IJ֒^>0\<~Ng_n Do~+ T]Ek5dﯮNJ(\rF|//&!dɁy.}.lER3 CaqDžys|y7{cd,牝˔ 1/i)=X=k1/1YڄT ;C~v7!" yF|O ^bPehA-o?ߌߣh7ӕ~؍+KJ} 0b?OJ&>Yfoqw1IaA6 ~λZVΦ}HBrD*_l ֽ jm9CVlA z͍"h\`McO-ݔ)fTPc%R!P$h%lt91҇hu,3;6SK1eף å'aWVco[O*zуFyaQ "Rr7gEzr1G"k%\λY 80XUA特},ަر5&% i^M9Tw~8'x5Nɝں־`s^f}-n| }6RVo/ۻE];/zy <*VoN֢XW ́f4w6xzΌȆNZE {:\MQ{z>4B[%֌~$O1k;d2H |O?PrZ}㜘tb|jxx0nW/5*n=ILs顠  >,w+HSjOe]o0lpIZ=N?2;LL珫~/T(H2 Q};y}1ǫÝc S+1?|.'9$yaNB h"!6rtp,).}DR"(lC=BtS if=j6Wf%_5b.F}jj#=o^3#E3UxT̛3U-Wj:pzj6X܁_27] nF(N鞕-)z~kRm9}zy0AVM~ |M;ZZI;LUG3UvfB1}6"xLGD,4l{ug&ȩz̞,s7Kx5 'T6=]RnG|_[k-Ghhng|4R0=A.vO ʃv; |(C=a;DnD(ȿdōȝ,l咆s YvL=%f˛wݲƆwԨ0:gO+NzZG{KWhZsNBaO!`CO~3jz?ٿ"AbOWU1JޕndF: ^GM נzd|wz)cTz"v*NH ze݋d.5(@65Yl4+v+N_YM~!c~gc2/r_iԖ,HW5cn?4(5*TOƉ4i*F{dDB}eNΏNPsHǕUv{8]>?^]TɝG%e@0g'xc $cΜc|i<|#% D$2HjHN;:/ O/֖"LN0% 3)(ƚ)\RKTIR -Hia?Nķ ?ã@|yD$FGXr| Ő4&ʇФ(hN ~Gɴ6,&=aHybr@4I1cy$S`*``aԙlRaw̕z?_]TI`j&Tb!%4g$$$ fL V^ҹ$1IQk=)+&X9 -(A NZyT8.5ApWsZ]Fζ}tH!h!~zjZéLmGb_m3uxhі!69tnv6i/2M9(`i;gWUf;*-CY'Hg+]+Dm){P.Ӝ 5Q0xw0J'ek(aE;)gXn?aE&(yxLxZ<}ՐԢ5t潟d%@8vCaaOPa\s_jwVb8^q"׈sZp󞅏D&vsOV6ڢFL6{d;WQƟgOq"RYBtyJ;@8U(DF7mm%Z-Q℡P0NVp .XUK% '8eSi?mc_!8%U}6 Y8S UVے91&)iċ9Ù!)KaYLw׽r0.H@]6FC0cR ġ155w_]9𧋩ÓZlOs^u |*]]o_Cව3BLݞ]: ׳؅?-EuCxz[z`f?olJ~Ry`ޘ/g!#_ޝP@?LNOw)$l8[ޙs&ӿ/6sw|Yີő15Nrc1Ѹ:&YwgYA vvU L2Z9JXF X"*HW )s2QSJ@2h2ſs%, %VVm(jAs49iSPx5v΢xʨwV*8*_mqPuZH,-4UIl hmpmS, ScmiVD-r7mкڸ y,ڊ䎝3Zw}0=Wt׫ +) a:\`!`XGd^X7dQ`4C(.Md Si~ΪGBOFz6?lxti*TkO_^q:YK[@]L d`e6bAgmr"VzYnE4o+TDx2GZ"ΚDrpd~m\QM*57ث0A'w}s#e[W*Qڠ#>m^gNWu(z3P:<8v7E$ rN[13_֓/qfQM65k:.YF3* >K*IN"%j#ȐN,JthL>?[RѬpSxq&4UvNd/5II+xʌY휌‘LyLƔyGxYR"ц9rgQJ6;XϘDJ*y(o_U@;BT m;ѣ7YybjoP Dzrf`{yhbld8ZL$}($AG#dQJhHfcͳ&(yߙ2@8R΂{oO&$9d%4ѥO>d^9DFA&#'<ǡFc͡r]Cnl_#(amAնuNx!?&feg@NG0 y⨽͠;!^=Ybjnf+6J+* yT*fEdr}}b$p'hB}lgJ˜̭`1ˈƃcKTi(@JNL,KƺDz[e2h }%?Am^G?[ԅ?+Yi"k^g5O<$ݑ?Hoň4\}9[>yLfCG\\%ZmQ\e| Yd`.//׽c(aZ E"}"% s&Ύj< dPdpw4Tkp^j `d\sܮS"8>eZh$LnJ$oW;xeeb i}0බSG4!qy8=;e lfg54h9ekDm8P씉'FN u6FR@ΦW LT`6, k 4@c8{͈]\1beK1$8l㍵5!$k̸/bD<)/,巐 [:Ek&4yk2^3BZ_HID~o7>d0d@nKr/_i S ģd^W. s߮}tՄM {_uCS3*V0|}+w*?(rbߞ|nw|Q?IY8Mg}M oPXZu A"~IikYIYrʪ^]œ'Gm aKm*^XᦋgBÀ܊J2W,<-Z+AMn+ͦ8먦ŕ;O{zǡ.^kj9M!4#ݴc=ǽQqǞ@ۻKJٖNu.iCt |RC ޵57ncn -/]SNfRM%.ڶ%$;ݳ%ːDZtj|xppp.f@P%NOw]=K1oo^ qŭ?>Q$ɳ ")"U O1rZ:>O+r>'ҌI/ǿ .$r m~8Vupuj;2}ere$aw\eVLEѳl?BrL*9~6ӻя+=i^gRV%3=cgIa^o_6>BIJ3T!e#H 9`<|#HČ'XNJIa((a"Ѫ)W>OD:pa2|܄N&C:Ð1n>A)2 -M'Ohx$mNiG[;(눀p]JʌƎvk cs8I}a9N) &a˪֮[joU&PJ S3of\,0q(g͔()QrMU¾̠pYFcH4hhGcC4o%R Óg5y3ĵ&`2)ay3 <A U+R+`BAb7]'U'`}o~)¤~NeH[jDP`7 fy36taxSd#vØYLvhVVza;i]H֦$O|_пlGN5_FXˋ o}n^ja2~0T:W4Zǫ^Z1 Ж}ZZMIQoO80~S [Լ3vR*E(!P$M5OQIG<9|r=AqϦ6S!aPWNS_o}<7zB9pZ)r; o d.GVPτ̙j4\ŪO )sޜ'z#3- I(4!|}e@@q/-fvu/l?kfF}g=8?>YzmgxSyqU,FCk.דs.򯂬L "NwKCfFwD8Md5aa".Y{)bD'ܜ)Ms)HZM\G/>d aaXz͹Qanٗe%w5}\8lvb16cT[ 9'pH̸"Ju'=aipaֻ#6*  q%ά zN@) *Jf+9E eNg yAO ҩB ŬQ+ 9bq)cBZC{`%ca-qsb@Jb0%95-Ԭ0]HxdH`K8'{E/f>fJd"Ut3mQe9Wf$ D%B*\}Ym= da2F2YօY d=$FRS)f6mU&H0UF]c9ӏtia$b8$%Ix<2<{hj_|YV>)Ey_57UrT!?- A|,3HO ktu!ǐŇ|i`$aOy ;)Q^˻*Jw6 h!NaM^/ 0H)q^@ d/\ti۬rA0izjTҳAзۯ\e6G0n~mjPi濤OeFFh`4:M:4+dk%sY?] +"NɡHHZݺsJ[T3J\#0u-HۨmN(0hkk9u{A2*s-؊{ oAiX<:ŭʵV4؊1E "'PFȱ<^n^;b۳Qz󼴻,银Ef4ov0PZп.*BF'{YTX=Kǖpp^/z#ikp: 8}\Fr9ēJ{ZOFP :Kit1ְ LpMKZ9i% BQ$d K32O)2sc@j-aj[0^*:Wu|~?CKG NQ lG}?+(zڜ~e|f26^,>^/:ٟNo՟k˥2?}nM"ם{^Am }zY~dvj@+Q0)w2mGv[6uaemܝȌwȊRܴI֫&+u.@+/a$mj8p iΌwrpN! t%1j0=:5Dqnmż; 99fOi} !6CϤ_EnnwtE sT ەeݺs&Z PؑY󽹖Vӷ}2ΨGS5tfwϝQ/#"܃yyY*uݽNjx\ʖM+*0}<4lCed vffӓR3$oDA Z_WWәҿ-_ f=kN/1>T&o21xD͸`5#lxg-A-#iF`ɶ GVbvS  4iAiJ !ȲCEQDL*`a{%R߃,p7sHj1@mzo+ Yfp f0R:YLű,({\pO. 8ZjZy:f@PyH j- 9Dj̠)43br(Ҝ #K"\Jv NsL)) ($F *̬m5 2W $fm0sّSmvQnc7ĨcGұ e rx8,y5M̏EȒ_yMh8V)v@jA+$V&g ±W۪ 5 mP#D0ɤ湤iV3̌jLSQpND !;b9EBTT"ETL"$2B,ts" I'$҄FzY F#s(1UQ=8LV*2sA)Qbch}ESL),"eF%Ik#/2F a)%\d)sE*k=ArY?%~ X h J T22rK(3$F v[eݝ] (DjN簠FC*+j̱@<(Ӧ7̹(8 9V}p x @E.D0+03g4XNPٲ:EFh&@|fpм=z!ȥri=c̏ZB5O 6S90Ϙ}#=ЇB] #_~^ bi67rʷ(;9soLw-&+j`u)ر m' &qq,(M&a7c5욘د?pC)<fa'K[vi4nTr?ͬp4 7"b;]JvMƛ"{y.?ǐқ\vZ~[Cݝ'^/{i6]V2=heD4j@`t_;d`1fUe,yAf(./ꕚM-_P41%"Mg1>zűi;_z΢˿>2/k4G_K> 6% %郆9+惧fYU{S_[68 ƙ5G`'v2vmeyUlex5K[y2S?*1mA}@о`E~ȬH/Ni9{NFZemV .)rgRKAm{*'-oݤ]nw/Es`mzfwNK$vعOӪsv9ɇ(=D>G1RоMK~(}dC#=3>?j)b3<"F-^-{:o'9gYozev>^KS-u{:n y;BsU"s~sY{:9M{{1@$W]bݺs:xECF'ϽjVF͸‰#{@ҦKk j򡝲|M/P$eNz &lshJVڅd'Ы^^gN2R>״뒠w?^]ngpH4)!R;.)Q *qQ{$d?{O+R[RNE17UwRq鐓S'PAp)F6:_bH/b6kې! 탌# >ۛKv8:it5 \¾!~U CmNU؜ѭ8h ]{RmECxA ?/۾ұ(vHƲ!-݊D2fk?_|sYTJJ9Y|%p,YӳT zL[Ie dG@U^2&MhK.|}ˀЪ#/kAzP{N&vq}І2OHi=L{K3)3Rӌ,Jkl}qN?cIT^LomQj:LL*Q0 L&:TPEVa^CD>,u˃DcNIGUX ׋TgɊݱrZ5/N!#1#ͳ Sre,9!Sur5_V5s$1h9qL`6b1/:j#'Ks{qR9qsu͟Px(ܲ8Uhϑ yd>h3-Qds~B;_KMyVYsԥpZQfvgȈp]:3R-Z9sthͨs Z?G68vذ8k{*>]i[4tI?O<|u+ ZiQ`+F巣B@Cqp0n*'GBZîB Cqpt Jǥr<৸uZ$9* r:1`Gby=稀QvU7*,ݔݍ 9jkXLx]ņUMÀ[y$Z44:0OO==utd #Zpֹ<  zl@QiX҈I$ k,H# 18<Aq,6[HG3x y曟X4#x]Sx +x #֐x叚psbi]# D4-}5#!akNFFKL#A0}04F#F Hޝ14B}i΂=/h[و 7v$٘bb6A/yētiLmf$860'|o1?1.x@OP,Gͧ1+e!4t`4ر,i`ٻ6dW>%XSꮾJ8޵'e}"/,OF ٳ$'ۀp.CJGJVG;ڔDmM|e,w\W8Tֶ;jW8)1V8o<jgF0BUT63ExAd $CA G>@c&|V^Qˣ9BirL1m Zg%QK<*AZY1 GJN"u{Mq{I"P&ZҋbPHRyZQ8v81RHkc?H-tat`~֊Q{q]6vjvkv؍OD&%9Icym7L9HEoh ޡf䆇8(G8h^_xj$;Y^ J,/#Orbdy1dy1dy1dy1Ӧ9C-x1g/?wj ˩bN;NsßsG{|BԢǻg,?xS5%HjO6&RlL"5lL"Z;lLj;t4&sȸ4#ᔢ4jnՆs)ES \Yd4xB$JI'6h'hO6& OߓZnTyq}HzZ\].p&FJ _FeѤSxyE_;g-C] DPW 0|Sw#0aZo.OoS~MS,P|swk׫ 큷|lt5? ,KcǶBqn'6y85([!T[yś8 [qASւgOF[)MTxzpNAy6ޭ 9q-)kb98w+Šcv<"4i)iɻ7 6nmhȉhNھǚ$?ȴ&Szm18ϻ7nmhȉhNzĻ)J1xR N=n9boޭٿwkCCN\EKt (=^M6_?w+ŠcvS޼[< [r*ZSƎgwӂcn4Q{lV-yz64UDy>nyȻa,Qvn9Iw+}$qޭ\CN\EKtA~(([)MTx\ߛw+辡wkCCN\EKtJ愊}mZf-J {V-A ҵgNX/˶i 3슲}S譤 @wm_Zk]k= S٢-?R=Zk{)*@i3kLԚoh םۻJ)X_ckmZ;u}1i ^Y }1j F՘ 156-ADjC4k}MK0T՘1}1j ՘#;k'Qc6H:Wc”k}EK44ٹ3-d_ck̭Z6s5fט۴4\WMטs\@jטsLw hטs1#J\5ܪ%]֘o"s_cn8Dk5}1j J@j̐}1i !k̸B_ck̭Z{5f%k}MKP3hD_ckmZʻWch_ck̭Z{5fC[4)8@gG#`7L0jwo.~>LDqo@Ĕ3#:rf䴈#%b'}dK m1 Cb# }s4)\@1xRa2d+\\NyW8cR寠dW3@_Myz5=j闛QZ2.ݨ]hv5ŭY bb.-.Fkڙ7]~^>EVeuf|vU 6'd^aKQ bG"l(?B(r`#ój<M>f>~ϟ5Hf{2Gwo&v\,< rc<lcزĸju[A(QJ?x86LwPp- }ݯF?zҧZVk7rn[gho]~&߱9{V'{7*ekՒ/ ?y ~&~dj8Hbq;V׸)Uw\Mw~ZYd#RIۃ Wd8-3ͯy k^Kt IŮ5D!A&Cczq"#)QcE4.,mx,5G{6ӓ$(E. Ծ,>Mf3maXH? +.aXjǪ]\ +z)2ḥ7`׮WQeUhLdfĉC (F" |ᩳ_~Z*EM U*Dn(>ګ6/"ߺ: aqXο yMZ+&a[BW¤3\X/K03䮖V(p寔-J jdVF(%C bP"i>i9,?#Y/x վ0fbqZDH ! )τd\ @Tz$aTe~kee"D]IEZkDZqZ[d& yBw(b?. ܒl&.ZX@j6KsC1"ƠkꦓU: =o& "$14>ݮԥ[9kok ko>{{HY1kj7ewE`} p /ީ0}3/f|0'ۤH+{ *E=k|oIr_$Iiidz8[!6? 7[Cf-tl*x1b[nO}lD䞗ko+v0:o7+U3 /7kɸlkڇN~qvuջ{('obW $f-ՄދOwm<=WEj`򷥾bQ@㇜.cqk_{6(*^ LJ`}gLh͜Nܠ6 fBKk H$/7k-zNbخ\quz?ig$qd%7V;4=V;vX"lJY7؍y_J+AТQ":̨6F9VGsev( b [7${Z0y (czm(ccDpex?(*t6q~Euy9{wqvz#b,D(T$쀄 GR,H-&HS_,B9ہ(k$# <(x`^VxJ9TW0iTSՎ>^'Qk<}bX.'2Yny\W[ķT6+o2͐:c!j'?\nڹ)%GU<$A-<ݎs)(9 r~BpԷ]*͠7_RU}  B|/ O$E&R4S}_}' ?Q6]un &+[Z7|| r 0sx' QN0Fa8 lQPa(0I4P2K ?2u J#0i$c 䱜yy2ٴ?;]dy[^Y03+ ;oˢ/{@!3=!-0j/XYK\kq<`V k'}ء ܢEH:{(C?-=_?r~!܎ } !2aNY\0EL$qO,OmKl H-@{|IH?֜#w %twj {&WBZs+!z%ܮHz 18L%;MK6hiI=DlTL֙uzS0{tkz~z3F&El0$B6 a?uCQ>sZJ>j)'&R^_)_f)A!(B-fÀ05\Gk"'_uTITk;a8p4 iUY2kwH8c0I=2D5IA<% hKɴu%w,UkoMТ) rYQߪDZ}=č>i^iHLrc/Ɖ)&6\:JF 8KV; aĂ"PQCTvTz<&T_l,leFTB %C:xc<*9jCڄ2: J13&0rV\l6&K X( lҕ TLg/]8j?m1͖9ͧ\ώ5W  k€.u, 4@8mɳDR-V_fyav]k6,%k̸ȕҗ>"@Fނ\&Μ&WF%B_faSƙ2˫ޒC:'tē2UY 8!m7 YNL. Zu۬۞IJ $Bufxfe!ܽڬ# ?lbf;mR"*ӮÆЫ%֝y3=Y4׿aN>z3G T $9rfx )qoZaH׺/Frm2cy˗1W aTC;'FMi[.Ãps梦ݔi!ܽ&R䨩;z$h5́sYӯo|Tǎd-&2t9EQN+qZv;ORCCKIՠ%[㴦M6u9JDzHla^,/_-E>ֹ_?Bt!>\h9N߷޹f}5~]>|ɼwgte=҇橿4;sE] rwwuRN?A NM[.WOzTsoۍ l^SRM(bv NO!N w?[,ʘ.vͫo/2 kvs4H.A5z~K6S;_'㗋*)wuNRMq8@ȝO"L[3 142 Xy)TLz<҂T|0WUEHo2*ThoCSYɸbƗgŇB:劫e-!ٿIO1D.Q?)]F(sI`9*9Psp"0$Ktddmh*3R̄v-UlR~M8{"muq#oj1Bzա)DDkLA#TxDE)-be ؋Wguڣ s0șiZWs=hW+ ]&C$?Y})іg9AmsAXT#R5 $s__eZ 8HI!p[C"T|oy5K(yt6J❈hkt2c$%dX29-9`r9/rQg/F2ȀV3mw,A'eس)^  SA H/߲P$x ǛL$YjuPLRϭ򲵫:,yԪV>j,Ta$Nѿ7 ☼ >Ι7G B9AT)]2KKE5U_Ë|I8/dq~*ԘvL否~!`:B,0Ulle3ZcK)C M^IN>0nc۳˃N!F;*`ogF$j 1*t#vGZLEQҨ ZR![((G+mƅ()ޑ 58i)t d(=( qC}$SR6Y=кg5B>óf5l$ɍYٳۉCy𿊋Fs,?*~ i|aw_؝vv_`E/U P +Rʔ1UqZ)½yy2y\ $wIl}u9aE_]zp^ _V5'teћl2sًHZ-rdZGHY)JwzqFhZ5x7:.5!TGq (%L)d Dt8Ι:itN3uZuF2-eNSJ!c! %. FXMi](4.~fL鶆;.ĭ)F3ԥJP[),0)fTZF8Y"`d D|{J4,pNTW{75*3!ٙQP0 H DQ4pc94yNb'Rp@,2 8Qr0v >rJT `%GIcH`t=h00:kt(uY\V?(ubfס1S=t20z=7hk\)q%+W± Y.puV i+$l=1ꇅrXYKss-=Cpyuذ|ɜ~* eښ6_a!g#Rs[Xʉ8JڊS,`0-Qo H@  3=t4C@ jab`3Ã@ ÑE0<:2]ZJR| _e`^5x2R5&+; ~IXygmࣄe94Ӫ|_}yn@ C"vu$'g'k?QT%ᔓJ _e/#K#?c)\d+&jyȥ%ҥVI_ϟor+P,{DM]Ê_šUgOm=X;`jRpvGf+e>3񝥗B"~8gOuqLQXVdvI8[po5u@AI…;G`vw][C)\h̋l 9W;6{c;E"ȭt(AgE5Gq?h ƽ|^f.}^OAH0v\*]0E0:Êՙ 0|4c4t.{(YGvTH)0U䅬JaHPԁR |a9teJSqOprW y{Ml*Pyi =2hr] WRCI┚ { /_zɧj2øہ51$S<͸vڛ^oyugr7 ?vי@"]6HP1Krx 9WvOشsKF++ 9,5D;|6~V~w tzafٵ7K 2GgNm~&sΡC ±R8\*M"T!0D zCjeD3*mkhV2F>9S>1'< g-p}N>̠q=?Fh%xTxt_=cBUŽV:)-B-B#J4=B12:zUQHٽBJcDyE !._"cATKJ;/@-MiSbrEhC+Ćܰ-LqAdDss_zθV߉z5JZZf0i%ދbi0:1XQCwe镡[ 8Չtʸǥ^YVvcpz}ւw.•x @/JS!sҝ| /ĦV/JBxajC/tG ='m~`Z*N^PeN}2r~ґ;'}Ź5GB$s"+嶛/cwt!Gs#3lCɞ.bKĪGЊκ+6)U ~<d5M!)!.E\3vǂ:+(𔂮 Huz't.؎2}i!RUuF{H+(`]e@+آ*umLe i,yVA DqD+AK7 \s`%͆r/e{vwRmϸFj:p+[ͻy YGźK5;ݼd;WXbً@P\Jlei/o-T@SSu`>&= `#mń{}KuGW0+ѵ(Whhzǹ\RV5WVi9Ɗ*[fQpQxEb7:Uv1*UcC;N;JCQ*( LJt1=7]䄀74{M[AnXH)ʞ&h_ii*E?cTbqD/n!3KÈ +aCEeq\+{c_Qk{p{ ,h=Ab%a {'+b`wX˧/W(j|N?Z'C؍?L"[N"[N"[N"[I6`~\zBy·V Ǒg/,D1#ӡT:J|:Êՙ["e`H1`ޅ1OMe|E,bd"l]!OMd]ܑcVZ4հ ? JdJ UbR^יt :h$)J x\Rè%Z\D渀DK$\9BTh9]o#Ke*S!X&UH(EVVx` NP g\ק\R:ww^7ΕP-꽯_Z#Х[ctK:Dݓ:1^ EH'C (̘,e_CSh\z qq(߅G>/8`qw}$ΨRuXv:5&a=jqn„GׄG|0 qWTa:nW㚁#QR}cː*̣ . Y;{oWIx wQwSţ8_ϗ<[ =xaww]NyӅbϨ}:0䟸f͸گBnif_ǎg}'q]{4>4vgqdS6lT%LE &hk0.wA۟2khhl[5@V7vV=*ݨ?yzH܅op>NDch>Ll3MxS76l~w9[;o}l~Ƨo*{3 ۀ=|;'xf|y~u{yKl4I0nrIn~l]~6>aԻzA\|,̈́ ~1ƛwa\z}vڢ=;ɋ)=@qԜ}NC ߶ⷷH`Ҭ=>bju|nXq:oƕKݝE܎wsKc ͎0F㵻N .jG`;m9t_V.<D]y۷X w6hsm7z5pMz*5# 2~!H3VK?C>;y#0'bO;װ?D+MRc(I`p>#yCg;ZEiK/F q*օ Iu%eeϬD)y1ݓLq8E ʒ};fmO;;|)qLJr*%E.,H=#9hm 9ڟN"h|:b))3(xgv'tU*<ūl x Ƃ0|=Oxf m#6Rm#6R+gpi(_ Pah-'53shz 3_r d/QlH(•%E[fZڗԳW?pN~qtӧO蹇Iqk# 40ڙbbv?di.tv.-ǔ(J`Ф,sHCf5"BY(j!gB !pE5QcAj#Ľ)•5,%uqw<`)q{)ّ>ds"\hEC8#d\4%G k[U,_824 !0؂rpR<434LN)F Œۺw ͨ`P$!JUkƒmJ_S?+ 6.B+9@bcoc;s{琠=מ.B=K!2Շp_gpD5@sܖPZs91=m: 챀{|rS W Piu@AN !$˙Q}5س!pu٫8'w-tњP n~sǛOsK"~RQP}k{1 %a\ۘRSSFͧ{0&5jPn\27y:^?*Gؐv_R'xEQhPx$&N:=SPa#*4<ҽEIdOs?I6)5=7?/{8u<u7QF/>S5ûv9e2dhsf2wg:ź6YZ!+;c$' B$/$L^&{1Oy~۠8bL &.荬N<,Q>MɮÐ`Th4*B̌, iC6"P=tUC+CE]SdH2uNK#)o> (9\X :m$clL`ZQ5/>OSO ?.ԺNjZQSE2F6':RV ɡ(76Ba5PlVR8V4iI"]Zѽ"T4d I*NR&H ͣ+rZ6 9`;mD)ΥʕDfS2cU #ITqKKF %Ѓ(P[ 夢3Zg*D,. 2NԠ RH  u >(|ؘSΔu.W2ϵURpUD8 %4-sE,תn`TZ7ò`ȘYPjP88rXQ<.g,J#2av0 , )ʢv,9/|Z#c8W0BAhᙹoo @9)2| , ji)Ff֖,NF-I[2z ժ\ {)hro 7rNePSLtn!6<[6Xf(:+h9w@].!.~% %Xrל`IA02J<+5x֜:]hZHOZ,N43Y1d@6@c_ ܒ!!@J{7CeQUb 󭡵5"@>z_a-mՐCb_&mr4D!4/P%\sZjTV;^!mj^1I[#%w2ޞ6N[!l'93^Gt"C*64tpji&ώ ?3ƱŔLVckHGZZю &-tRzEi@Qd:<1yIMȊLM BQMUх~a44ddS6_&Gt_q2ܤ" 7+1]Ցř0,q'!:B5bYu]<+~dx_.W!ه[6s?sX^zߜׇWmo/7$'F/5aɫDS \\WjxROw+S3 ;>eY9~ZG)9<;߬}WeeXBAf@Yjb RQl97sm/yPmVDن=T ùl>{ZV )bq񃪠S4uRȊaPJħi~4uDFR:/;U=ţ,b2T7Y}$IhCu!Ih+sCU.V,ɁYY -%5(%UƴՅvJ{7tM4>, ;&h_;]ݪUbBHR9j8z548#8Ғ[j(H`\.$+q@‹4'P 4D;n]I ȏ% Iip|KʲS m]FFܨk]6˲PKrIya1Ѕ 䎀ӄk UHT3!1RAs8(ˉTqJDv@f9Q'̹YY&BPY)TW!^|X!ɖ=bo qVaߜjo~ٕ9Y`g1k=g{ Su@jXEiKx^w}d $SOU%,4? PR`L T%FgX0%j@ML5a,n%4?/ּS:&oix`13m%Q+tlaÿ|fZ:`)v XNBIWJXE%gOKAbPR~ٕpjU<ܬ~.U}濽9uuOr@ٹ,ןoapzɔVⷳ7wtX˷j\w8חR%~R_,wRKݬ\#_ξ b-]~մgןo~^ y<\vd%QܑKMƋG;p׆U\(|8($ZQ{0Y7ofEb#|ázJD䃝#E4=F"60 %i1T I= CL}y5N~vg/Oj}7^Fc^]MYMSVF+Ll,|lh>X%FȫD6s5nTOL]^څ|wIՈQYI<|5U{|& qoD2fd=yFU뇎i"Fı=ı7Jvq=Su{3R>dvPːFrF8cB^N7W;J74!Uh&z6GnP&K8 |]>7"XW*X?;*WA8olvfd&NFƃS␘HrL 14/)N /)$J-$*)GIᡦz%Gm*iqibH[L}#{{7@^Cim(B! RSp**G.ҕ*DY+@,U`\XlŀRReL[]h7EHм]vhqnkj(od{bǹ?{{p7+oo>gBX|]ܯվ?SQ!󛇫{ٻ޶qeW|;{/Ҥ{Pnќ"kbcZr9/))ر$U C[y89o2Ǡg gd9˃yoY%N31(Fkʭlj픲,tnT^7Y-5A~|O=^ݘ ɆaR2Ά_3#o>D)%` 4GVS -J -)j4$bhHJ-3Ӑ4$)=je#AY^9oGTӫ(QT8udǡa<]M1hq18V}$d@Tlds̥**yיox\-51~R?(/nt}- 3_-wK_7z*FUrL FB#A\V#,nAfL8*Jis]%eE,ZB!dōhy~3Z(tXMڟ~t4*=3~)šX#@i-Oԯw[Z>~3C&Pq&ӹN͠T:bU#9qT{nн$zҫ|>rdͳ Wx'JEX *A}r uew[T2=P c)o >:ʡںY*x&͎sG9=U?_la)_v c {!dy0 >cK > P3$ke70AԽRnMZ !2i52 bu4߼토Aa5:r ᄘ)jH⃭}CּQzI&KJ Ȝ߆~! .u'qwM9a/Cޮ&=2uMc$0ѐ# z!SƬ 2@~Q +O3\VpRa\5=;#$aȷ5g@In4,^T\Z:l-3Ff 5l/hs"PĜDE2 q`7Bu.Y%Ş!ŌJ.w]±=BcXF,BD =NE]~%H r &\ojwWɃL*2T.8NF*,4N }e2zD)Tj ߍnxsntйAّ?+!kgq u{QwБW{pHЫIt y F3}Ŗ|9Iv<"@' ʇyUZB9 6L_&-SFPc&SʘlإܴŅ9%;Z%1Bȵ|WFGY (s8*t42r~=>T.t(aMmHRNL(%TvJ}T01.m/i=T4vݛ|v/&=yQ"˞^*<{**fܲ'Kb{]<޽u* o\%s5l+)G,kr@%gFٗs|s+ NFlil9勆B \i!l"s[s!Jӈ8ԉM8/wg_щofZLsn|~z!-cpO?b[u]6\?؅,$_aGW ) NJx!>rt*EzC!i^8SSb=.ApnbZ+ #dai U*iZ6Ye#4uŒ渳&HuUnc;%T *9JVR%=8iaSO(mq=uz&ªrjȮZ=`U%[vX詭VuuwoU5)tV`U=tPK`U%ڈ;oM9q!ߤK"&'~Z?75 (EZUOļp/s\ap MT BzRi>d'Jڜäe fF|SR{AjJ PaQ/>BcLY!qQeC WW͘jdYn3[OpďCw?x&cpV[|BxAittBd\OCWb4_.ҟ7T"t*JʟyZJ(<'B/&pW77 6ђX7_ݙZ63`(`I}"iUO L»1r`t뭚aXZwu2WpZ#{ʌ=}Ya x0'4@pA9BR,Y\0)(r }3WM7K*(\_>#kM^D`DEI]~ѥų2Z9 W84ԖﮒTԏ*U.8NF*-'NYoe2z8Q 7FW? 䐟¹{Rwu{yd џHgɚF0i1ӗ͗iC|D 69YS!.Ѿt]#E=2,<|)sqa쥲~~NyAm Z+#: :&+=pSaűmo`{\Z=5Qj#ѳe}=Y&Y?դZF0rd5Yhj k<n' Ӹ#(wN{ ^@_4rt vkvEpތJ ܪHnS٪z(6VOۿVUҔof֢j(!մjU%70XZoV*[SuvpP,/(J㾾8nVy LOM,1> `%MS.&u58ZiQME3!GݬE#ZOM=ؾ'Imd;`S9"ctzrXIs*uMgs⧥3{c]cr]Z5JG ?'ɵH o&V++)7˵)mq ˍZMHjz^Z;LE{>JbH\d&HV(ZGyQ(d^[/~Vj/GΓ5T* }ڟ=ự9og gd9˃yoY%NcXLSe @lz30%&Ey}|FO69߇I8~͌?$<Ȯ2Ey!(~֑;2fI<_i3j$ȸ8(dq'^-$W)7GmEo(Ջ]hu $$X~t$u#;I#M&mQ_B_rzRXiR Tl6Fj]>[S;nL;ijZC^8e5zq]֍w1) @J AG/dk2KB5Ro$xTV#/ߖa6@)po5JڅF2Y0NgsFx^-GGco<]W2yQ/~NtSٳ^f34 4j"4&P|̨0]W`͹_x5M^!s4E,Ja۩9| mMW77 tv%lFX=3>BJD˩' UL٢sj@)/1,p FןުƚQ^>|H`J "nP{rHgm^#b\4UN"[a&kcXMڕ~@<)']d߹s]ځ)Q/!{& LXo;2HܥsϝACZ%F1 M[J2W=T}*lW@A)@獱$\_uӗ%[^jE8?oZ:N tjҘUbm5#r B%Y.ѠhSuǐCdRYdRjđ¶ͰVayx B&Z &_%c,r\ n&E vmAFp.rZ . gpY'JkWO#j!Ai&hC6)Uc}dʌpah̳ IbFq+SQ}D>ۚlmH!`+RUO7a ECҵi0da5dXM:]1f&pX !.J8 )zx(3c'Cf\607>BCmV۰ >gj #! @b$.P6ְ BJ``o⚗C()qd%2,ZT}"nqKLi\7"0<̱G. <][s"G+D =퍉x&k 7Z@}=YT $*@-mT*|皷/A2ϝ):ǹ+!xXziEs+]a$w0ApƼ>hMI X*UnS`ɡBY}ٛ>I?Cl 7_( KOr&N磴o8gJ΄8WZiC;}wl?̏؈[5A@0xv(?#%sӠE3X{sfn>^@hu3隆hlrkClS..-YGh,5Xx]QSc_3jQ,b՝~AogGwHfu: R;2cX=5/C̛%t >' *qGO5 )v{괩Z괩:7?z;eVNUsؕYΕyyv/g'.Sc[Itq[㈏;C_:>Ԛtmw4UpC*w{f<[TZ(炼 ]'8;#<t*Spkgd& ;ʼn77Lu d:'ߪ A̺Nnu ɓM~%pOAWM\ *be>P(,.!X/ƽe4fpNAhn݀@ ޮ+wV(NnwO/'ڎoҦrf b3yqM2of4\{K//Obj{~r39JvWc f3oy;- $KUf#D]A|r]1]01XoaP߫XnWFda= SACWT9[_;M6e(9R͈Hziѿ{ ߣ.S2of?Mח7]mhT$#w|oR|sRWLeLze/G~m4Q ˶1ib\y}rNu#5ac6zV\4ms?QY0O'zl78[㧴wN:iRP_KLAabp- 9~")&D0 Ξ~9A4ϕRGR8(0PlqU WHȦ3E5/Bm*0!F!~>X "`zhA|RIB'rK((Be>xN.oH5>1VJNdP>*T.8K(83/-DDuQMp%cGrjsHcL'6ABa,b9Ut*N1e_yl8q.#7;yTw 0baD-2va1cal˓8 - [_PXQA7=%W|Z`^pVr鱯rVٍ!Gl.tv !ED`w$p?a4@devuryhh58#/G3(SxwX&l %L\،"R `@^Y䒡:cn<䎦XT|':Bx( p  ߪN`h5ՋJF}Y2h)q%0Wph'84PO ?7.(CC`c Xo0寚9D_+Kxi]^f]c NN>xJ#/tsnB_ yH@-XuAX) +)l(Κy^ ͎^%eaci*կEZKe{&E`BVi0eNrxQsr.CR‡HQZ 2P.YoZ4FNvmw gF--ԔDbU_Jq_֗v蹕ϕCE>OE^Ze0ɤ"MqDg۞TzΈ&h]+ fK۞paUQ (/K G= F7\Nv\帄U RU=eQ+1a%'UUKi"j5~ikw2mˬaYNccynϖAQFH81暂Sjp>'<`VCܺxg ˥N-X<0V8g;g󹏰?uy;V>om-E{" eT-Pwĵ}rCwOց=ϟ2@,'`Cup7'b]s#>!WQ' KsNz O}D[Ǭ6iF}rBJsH|i+ǥ&aRDDT~h-3QH@x.Ĥ )ib7>0C Rb,4QKL<Ƥw` !(gj;(O.rّJyE2RufaCTP Q)1@*%&o>b(ez ffCIO+(zTʷQb噚rJf1jZk^gNA~ư;AotMuW̼xV$L/!&cagljȎz88n6/2qꯊuuu ޽rw"0[ |{,ZkE\ B+Lk@cQWOc8R1WE|U7+$c{/^ԇ5%w0̺)=Z`ⵞ;܏9}YRՇ8h?]~qіpN,9(VNb '0%eGR;j;6[{iLtLt-_~s 8•{l=E1n6 #T?;oXm2**2_&TVt_A;RMaF7'Pr3Evx2\%t\;JVr=E+!7~k iIbKyyRcJX(N|dSlp,ջO6;tܬݪz7WV׎[av1^>\|:oXϪ;WvqW8Յ%oq2*rMǁtj~eyvdjG^P-䅫| œذ~޹ϭ?WߥnD28zAڌ^wTp`KX_QEpC1^0pD-5*Wt|* CY]>2_qmΛ] uWV dz1 ZsMVje_qT1mOִֺbo} ~Vk}s=+W/mWi`3+3ҼZ޾Y]YU\B׮-~˛çE~woNKGRFeE3*N͋|&z=͉-MXvzH~xEJ!4};^ŵjEw~oWOvrMȐ=?R4FMFp$w(J~r Er'D)|(}dE ]d(M)<챖0D׬? 0bڕ+x݁ UY 9.ibn;OkMPЃwrK7J.\lKμ yޞjrߴ^W''.GO2jøhW+6_0FBOV`͗<_?MqVff67&G!{5)YAxUzZͳ̩r+lG%kwd@μ5{l&*%$ye-@;#-OZ pPrpV$Wdɲ%j9^v{ Bڡ7;XQ0Jչ @͂+"z!qZ#Y^M [Zi)S("`M dEZ_f2Q$`3"jbۄ C\ f@h od[Rbں2@ǤB6E k" FFD2o$ȶ cTP%pJל>fmDEUvsqaZ9ƣq `Y lA_F 4vs8+稨F<'"{`PtOx|MHۧ=19r=m!{sj1?^!,jxs`#;BM,o@A}vu9/м0Bƞ)ӗ0a8򫭎$P A:/ ;K V{\B[qN574ܝys*~|rF#ּ бl6(0wʫ@ bCsJN2gA)) ]cIdzK AL.(Kͨo!H&l Hm?6E7pp|'_nG[YNf3r'l*B 폡Ȱ DzyQf+ NZ@N'Pv˕TNll=vO1~t1+q7EwGuP׃' 7h\6hh@_rB7l l-"Q" an$jPA>#azOfg~ѐSQ XU ]A Ef C-pz~$Q%\J\#L ujWb'Ԩ{a{[)BtF%SLfA`Fر"{DVA^mi-Q7hUp%_/U]=('^ȩl8U "[lbJ(`\쥁URt̞P-ݠH~tn,Ai q—!w`|zܘ򉂀Ój@ F3v]rLj6-_$V5C&zs%[&Pʹ G?0fՔv, <6pT7'9ԃ޳XqJFKpKZ?޳E>8\]=81|YRkM ?ZH#yI dF3b#O5't!!/\DdM?>nI K DtRKx8*omvBB^&TTDڍInN3bx?SncO[E4Ij3NFhO5vT'PoObS=cvg{+ (yԠQ(ZN,X#9\rIeRdɶ[ Z 2b5Ӷ{ :XRyvC΂rpC"o1/>LQ=;4,`r!Au[섷XMzu[eVISsO).R.w%y&yέB:/c$AsÐy [3FZZ8rb%4aGB ^zsBUV_.~I~`{5A=V jL1=  "u=M>2?(յWK k*x_>doĂ K1@O=7oy9oMCFL -2Ҽ`rC'(Ӆ) P(.7ZX9t (P[p`R`ˆ)ҥ3ޫX.q;s;-E+sN^$F*º.SZT{/p[,rZdŒ`S/<<w3 9pA xf6tqhcEFG1Es-P#;\:8/Qr82Ferrx$"Eu~Pc8ϯTڎ`.Cvv}!fސlP =1yyOjPLAHR3b=޾ZLׅpMǼɇMI b":ψncYYطvK! n]H = ,.kwa?򕠠x\ |%|%dY~-_|%WqqǍiL JX .O6öq&poeUu>5O+dPSeA 4RPg))$GW ?L gGvoq5; h..|kjLϴ{D҆d)mdK# N%VI3 84UAtWT!X[JAbI!z0Óh3tiTyV(i5qEVA̅G\y y*kfeE),(Fgև$P̀E,TBN;^V/M2*W;ǎy-?3Anvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003355006215145725227017713 0ustar rootrootFeb 20 00:08:34 crc systemd[1]: Starting Kubernetes Kubelet... Feb 20 00:08:34 crc restorecon[4714]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:34 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 00:08:35 crc restorecon[4714]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 20 00:08:36 crc kubenswrapper[4750]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 00:08:36 crc kubenswrapper[4750]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 20 00:08:36 crc kubenswrapper[4750]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 00:08:36 crc kubenswrapper[4750]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 00:08:36 crc kubenswrapper[4750]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 20 00:08:36 crc kubenswrapper[4750]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.276143 4750 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281443 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281478 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281492 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281503 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281514 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281524 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281532 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281541 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281564 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281573 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281581 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281590 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281598 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281607 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281615 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281623 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281632 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281640 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281648 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281656 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281664 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281673 4750 feature_gate.go:330] unrecognized feature gate: Example Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281681 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281689 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281701 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281713 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281726 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281736 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281745 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281754 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281762 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281770 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281779 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281788 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281796 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281804 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281812 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281821 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281831 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281839 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281847 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281855 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281864 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281872 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281881 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281889 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281898 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281906 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281914 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281923 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281931 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281940 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281949 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281958 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281966 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281974 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281985 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.281993 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282002 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282010 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282018 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282026 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282034 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282043 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282051 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282059 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282067 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282075 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282086 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282094 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.282103 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283785 4750 flags.go:64] FLAG: --address="0.0.0.0" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283818 4750 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283839 4750 flags.go:64] FLAG: --anonymous-auth="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283851 4750 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283864 4750 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283874 4750 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283888 4750 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283900 4750 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283910 4750 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283921 4750 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283932 4750 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283943 4750 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283954 4750 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283966 4750 flags.go:64] FLAG: --cgroup-root="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283976 4750 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283986 4750 flags.go:64] FLAG: --client-ca-file="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.283995 4750 flags.go:64] FLAG: --cloud-config="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284005 4750 flags.go:64] FLAG: --cloud-provider="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284015 4750 flags.go:64] FLAG: --cluster-dns="[]" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284026 4750 flags.go:64] FLAG: --cluster-domain="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284035 4750 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284045 4750 flags.go:64] FLAG: --config-dir="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284054 4750 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284065 4750 flags.go:64] FLAG: --container-log-max-files="5" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284078 4750 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284087 4750 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284097 4750 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284107 4750 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284146 4750 flags.go:64] FLAG: --contention-profiling="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284156 4750 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284165 4750 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284175 4750 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284184 4750 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284197 4750 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284207 4750 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284217 4750 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284226 4750 flags.go:64] FLAG: --enable-load-reader="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284236 4750 flags.go:64] FLAG: --enable-server="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284245 4750 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284259 4750 flags.go:64] FLAG: --event-burst="100" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284269 4750 flags.go:64] FLAG: --event-qps="50" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284278 4750 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284298 4750 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284307 4750 flags.go:64] FLAG: --eviction-hard="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284319 4750 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284329 4750 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284339 4750 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284348 4750 flags.go:64] FLAG: --eviction-soft="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284358 4750 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284368 4750 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284377 4750 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284387 4750 flags.go:64] FLAG: --experimental-mounter-path="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284396 4750 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284406 4750 flags.go:64] FLAG: --fail-swap-on="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284416 4750 flags.go:64] FLAG: --feature-gates="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284427 4750 flags.go:64] FLAG: --file-check-frequency="20s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284439 4750 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284450 4750 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284460 4750 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284469 4750 flags.go:64] FLAG: --healthz-port="10248" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284480 4750 flags.go:64] FLAG: --help="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284490 4750 flags.go:64] FLAG: --hostname-override="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284499 4750 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284509 4750 flags.go:64] FLAG: --http-check-frequency="20s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284518 4750 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284528 4750 flags.go:64] FLAG: --image-credential-provider-config="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284538 4750 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284547 4750 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284557 4750 flags.go:64] FLAG: --image-service-endpoint="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284567 4750 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284577 4750 flags.go:64] FLAG: --kube-api-burst="100" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284586 4750 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284596 4750 flags.go:64] FLAG: --kube-api-qps="50" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284619 4750 flags.go:64] FLAG: --kube-reserved="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284632 4750 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284641 4750 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284652 4750 flags.go:64] FLAG: --kubelet-cgroups="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284661 4750 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284670 4750 flags.go:64] FLAG: --lock-file="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284680 4750 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284689 4750 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284701 4750 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284715 4750 flags.go:64] FLAG: --log-json-split-stream="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284725 4750 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284734 4750 flags.go:64] FLAG: --log-text-split-stream="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284744 4750 flags.go:64] FLAG: --logging-format="text" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284753 4750 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284764 4750 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284774 4750 flags.go:64] FLAG: --manifest-url="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284784 4750 flags.go:64] FLAG: --manifest-url-header="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284797 4750 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284807 4750 flags.go:64] FLAG: --max-open-files="1000000" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284819 4750 flags.go:64] FLAG: --max-pods="110" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284828 4750 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284838 4750 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284847 4750 flags.go:64] FLAG: --memory-manager-policy="None" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284857 4750 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284867 4750 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284877 4750 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284887 4750 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284908 4750 flags.go:64] FLAG: --node-status-max-images="50" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284918 4750 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284929 4750 flags.go:64] FLAG: --oom-score-adj="-999" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284939 4750 flags.go:64] FLAG: --pod-cidr="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284948 4750 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284961 4750 flags.go:64] FLAG: --pod-manifest-path="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284972 4750 flags.go:64] FLAG: --pod-max-pids="-1" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284985 4750 flags.go:64] FLAG: --pods-per-core="0" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.284994 4750 flags.go:64] FLAG: --port="10250" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285004 4750 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285014 4750 flags.go:64] FLAG: --provider-id="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285024 4750 flags.go:64] FLAG: --qos-reserved="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285033 4750 flags.go:64] FLAG: --read-only-port="10255" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285043 4750 flags.go:64] FLAG: --register-node="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285053 4750 flags.go:64] FLAG: --register-schedulable="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285062 4750 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285077 4750 flags.go:64] FLAG: --registry-burst="10" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285088 4750 flags.go:64] FLAG: --registry-qps="5" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285098 4750 flags.go:64] FLAG: --reserved-cpus="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285108 4750 flags.go:64] FLAG: --reserved-memory="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285146 4750 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285156 4750 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285166 4750 flags.go:64] FLAG: --rotate-certificates="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285176 4750 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285185 4750 flags.go:64] FLAG: --runonce="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285195 4750 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285205 4750 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285215 4750 flags.go:64] FLAG: --seccomp-default="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285224 4750 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285234 4750 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285243 4750 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285253 4750 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285263 4750 flags.go:64] FLAG: --storage-driver-password="root" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285273 4750 flags.go:64] FLAG: --storage-driver-secure="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285282 4750 flags.go:64] FLAG: --storage-driver-table="stats" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285292 4750 flags.go:64] FLAG: --storage-driver-user="root" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285301 4750 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285311 4750 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285321 4750 flags.go:64] FLAG: --system-cgroups="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285335 4750 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285349 4750 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285359 4750 flags.go:64] FLAG: --tls-cert-file="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285368 4750 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285379 4750 flags.go:64] FLAG: --tls-min-version="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285389 4750 flags.go:64] FLAG: --tls-private-key-file="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285398 4750 flags.go:64] FLAG: --topology-manager-policy="none" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285408 4750 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285417 4750 flags.go:64] FLAG: --topology-manager-scope="container" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285428 4750 flags.go:64] FLAG: --v="2" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285440 4750 flags.go:64] FLAG: --version="false" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285452 4750 flags.go:64] FLAG: --vmodule="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285464 4750 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.285474 4750 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285688 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285699 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285708 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285717 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285726 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285735 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285743 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285752 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285761 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285769 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285778 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285786 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285795 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285803 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285811 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285820 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285828 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285836 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285848 4750 feature_gate.go:330] unrecognized feature gate: Example Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285857 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285865 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285878 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285889 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285899 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285910 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285919 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285929 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285937 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285945 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285954 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285967 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285976 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285985 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.285994 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286002 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286010 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286023 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286032 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286041 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286049 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286057 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286066 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286077 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286089 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286100 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286111 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286145 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286155 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286163 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286172 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286183 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286191 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286203 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286213 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286222 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286231 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286241 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286250 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286259 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286267 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286275 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286283 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286295 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286303 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286312 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286320 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286328 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286336 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286345 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286353 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.286361 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.287250 4750 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.299930 4750 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.299978 4750 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300098 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300112 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300147 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300157 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300167 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300176 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300185 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300194 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300203 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300212 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300220 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300230 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300238 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300245 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300254 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300262 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300270 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300277 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300285 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300294 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300302 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300310 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300321 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300330 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300338 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300347 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300355 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300366 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300378 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300387 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300396 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300404 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300414 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300424 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300432 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300442 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300452 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300465 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300473 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300481 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300489 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300497 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300505 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300512 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300520 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300528 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300536 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300544 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300552 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300560 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300567 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300575 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300583 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300591 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300599 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300606 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300616 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300624 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300632 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300639 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300647 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300656 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300664 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300671 4750 feature_gate.go:330] unrecognized feature gate: Example Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300679 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300686 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300694 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300702 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300709 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300718 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300726 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.300739 4750 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300951 4750 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300963 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300973 4750 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300982 4750 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300991 4750 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.300999 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301007 4750 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301015 4750 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301025 4750 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301034 4750 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301042 4750 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301052 4750 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301059 4750 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301069 4750 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301081 4750 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301090 4750 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301098 4750 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301107 4750 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301139 4750 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301148 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301156 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301166 4750 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301177 4750 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301187 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301195 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301203 4750 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301211 4750 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301219 4750 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301227 4750 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301234 4750 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301243 4750 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301251 4750 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301258 4750 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301266 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301273 4750 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301281 4750 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301289 4750 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301296 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301304 4750 feature_gate.go:330] unrecognized feature gate: Example Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301311 4750 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301319 4750 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301327 4750 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301334 4750 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301342 4750 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301349 4750 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301357 4750 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301365 4750 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301374 4750 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301382 4750 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301390 4750 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301397 4750 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301405 4750 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301413 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301420 4750 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301431 4750 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301440 4750 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301448 4750 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301456 4750 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301463 4750 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301471 4750 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301481 4750 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301490 4750 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301501 4750 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301510 4750 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301519 4750 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301529 4750 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301537 4750 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301546 4750 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301554 4750 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301562 4750 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.301569 4750 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.301581 4750 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.301839 4750 server.go:940] "Client rotation is on, will bootstrap in background" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.311462 4750 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.311567 4750 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.313378 4750 server.go:997] "Starting client certificate rotation" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.313409 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.313696 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-06 01:24:03.802684794 +0000 UTC Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.313826 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.341294 4750 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.343720 4750 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.346978 4750 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.364997 4750 log.go:25] "Validated CRI v1 runtime API" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.395212 4750 log.go:25] "Validated CRI v1 image API" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.397744 4750 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.404303 4750 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-20-00-00-51-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.404356 4750 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.437589 4750 manager.go:217] Machine: {Timestamp:2026-02-20 00:08:36.433247851 +0000 UTC m=+0.628083960 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:867946a5-592b-4cd2-93c8-f70a12380801 BootID:d9107560-2b75-4572-807e-08297e5eaea6 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:04:32:50 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:04:32:50 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:08:42:30 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:3a:5d:f1 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:21:1d:9b Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d4:b0:a6 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1e:81:12:d9:2a:6f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:4a:43:8e:c7:7a:4b Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.438019 4750 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.438361 4750 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.438992 4750 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.439332 4750 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.439397 4750 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.439707 4750 topology_manager.go:138] "Creating topology manager with none policy" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.439725 4750 container_manager_linux.go:303] "Creating device plugin manager" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.440408 4750 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.440467 4750 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.441353 4750 state_mem.go:36] "Initialized new in-memory state store" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.441488 4750 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.445926 4750 kubelet.go:418] "Attempting to sync node with API server" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.445961 4750 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.445985 4750 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.446006 4750 kubelet.go:324] "Adding apiserver pod source" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.446024 4750 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.452756 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.452858 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.452984 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.452920 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.453522 4750 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.455685 4750 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.458093 4750 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459631 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459673 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459687 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459700 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459722 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459735 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459748 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459769 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459785 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459800 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459817 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.459830 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.462217 4750 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.462796 4750 server.go:1280] "Started kubelet" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.464492 4750 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.464501 4750 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.464590 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:36 crc systemd[1]: Started Kubernetes Kubelet. Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.465662 4750 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.466267 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.466320 4750 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.466540 4750 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.466580 4750 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.466775 4750 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.466542 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.466549 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 14:44:29.8933448 +0000 UTC Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.468569 4750 factory.go:55] Registering systemd factory Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.468590 4750 factory.go:221] Registration of the systemd container factory successfully Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.468543 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.468930 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.469176 4750 factory.go:153] Registering CRI-O factory Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.469205 4750 factory.go:221] Registration of the crio container factory successfully Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.469314 4750 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.469350 4750 factory.go:103] Registering Raw factory Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.469389 4750 manager.go:1196] Started watching for new ooms in manager Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.470505 4750 manager.go:319] Starting recovery of all containers Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.470952 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="200ms" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.472349 4750 server.go:460] "Adding debug handlers to kubelet server" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490343 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490420 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490443 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490462 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490484 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490504 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490524 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490545 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490567 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490586 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490605 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490624 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490642 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490664 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490682 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490702 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490728 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490752 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490774 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490797 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490824 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490846 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490869 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490925 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.490999 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491027 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491058 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491087 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491185 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491211 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491236 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491263 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491290 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491315 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491338 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491371 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491396 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491421 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491447 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491472 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491496 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.490325 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1895cbce8efc4875 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 00:08:36.462758005 +0000 UTC m=+0.657594094,LastTimestamp:2026-02-20 00:08:36.462758005 +0000 UTC m=+0.657594094,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491522 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491601 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491640 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491664 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491687 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491706 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491727 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491748 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491768 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491789 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491810 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491839 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491863 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491885 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491908 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491928 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491948 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491968 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.491989 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492008 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492028 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492049 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492068 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492088 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492107 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492181 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492202 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492224 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492244 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492264 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492285 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492304 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492325 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492346 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492367 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492386 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492406 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492426 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492445 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492466 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492513 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492533 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492554 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492574 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492594 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492614 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492635 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492656 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492677 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492697 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492720 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492740 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492762 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492785 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492806 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492826 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492846 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492865 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492886 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492908 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492928 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492951 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.492972 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493000 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493023 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493047 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493070 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493092 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493138 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493162 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493192 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493214 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493238 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493259 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493280 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493301 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493322 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493341 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493361 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493385 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.493404 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495476 4750 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495520 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495544 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495565 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495598 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495618 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495637 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495656 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495677 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495715 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495735 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495754 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495773 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495791 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495809 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495838 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495856 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495877 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495894 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495922 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495941 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495968 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.495989 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496006 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496024 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496042 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496061 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496079 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496098 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496142 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496165 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496185 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496207 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496229 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496259 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496277 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496299 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496318 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496338 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496359 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496381 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496401 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496422 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496442 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496462 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496481 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496501 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496523 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496543 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496564 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496585 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496605 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496625 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496645 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496671 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496692 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496712 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496733 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496753 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496774 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496794 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496815 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496834 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496854 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496874 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496895 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496913 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496934 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.496955 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497009 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497036 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497092 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497110 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497186 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497205 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497225 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497244 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497263 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497282 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497303 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497322 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497342 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497362 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497381 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497400 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497420 4750 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497443 4750 reconstruct.go:97] "Volume reconstruction finished" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.497457 4750 reconciler.go:26] "Reconciler: start to sync state" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.515986 4750 manager.go:324] Recovery completed Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.528170 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.529798 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.529849 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.529866 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.530865 4750 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.530880 4750 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.530932 4750 state_mem.go:36] "Initialized new in-memory state store" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.548387 4750 policy_none.go:49] "None policy: Start" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.551941 4750 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.551988 4750 state_mem.go:35] "Initializing new in-memory state store" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.555416 4750 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.558547 4750 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.558579 4750 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.558608 4750 kubelet.go:2335] "Starting kubelet main sync loop" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.558651 4750 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 20 00:08:36 crc kubenswrapper[4750]: W0220 00:08:36.560412 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.560486 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.566852 4750 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.622593 4750 manager.go:334] "Starting Device Plugin manager" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.622682 4750 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.622702 4750 server.go:79] "Starting device plugin registration server" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.623480 4750 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.623514 4750 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.623674 4750 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.623934 4750 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.623949 4750 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.638910 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.659340 4750 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.659511 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.660995 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.661042 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.661058 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.661243 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.661504 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.661547 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662287 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662342 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662360 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662399 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662440 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662460 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662504 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662711 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.662776 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664258 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664299 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664316 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664365 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664420 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664436 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664588 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664727 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.664764 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.665789 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.665818 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.665795 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.665835 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.665852 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.665869 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.666015 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.666238 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.666299 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667100 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667174 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667191 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667566 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667621 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667629 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667661 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.667679 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.668618 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.668649 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.668664 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.672017 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="400ms" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700267 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700309 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700336 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700358 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700383 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700404 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700424 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700445 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700466 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700486 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700506 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700550 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700571 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700596 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.700616 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.724096 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.725666 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.725717 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.725735 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.725765 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.726332 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.801829 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.801917 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.801964 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.801999 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802034 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802068 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802100 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802154 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802183 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802260 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802296 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802298 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802411 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802466 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802428 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802431 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802543 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802565 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802596 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802551 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802667 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802701 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802734 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802764 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802793 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802901 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802908 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.802977 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.927455 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.928994 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.929025 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.929035 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.929055 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 00:08:36 crc kubenswrapper[4750]: E0220 00:08:36.929527 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Feb 20 00:08:36 crc kubenswrapper[4750]: I0220 00:08:36.996014 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.003532 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.024142 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.039353 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-5ca0e5998df1ce6cbde43b1a73267f375e787857bdd8a0915997cd70121c3d1f WatchSource:0}: Error finding container 5ca0e5998df1ce6cbde43b1a73267f375e787857bdd8a0915997cd70121c3d1f: Status 404 returned error can't find the container with id 5ca0e5998df1ce6cbde43b1a73267f375e787857bdd8a0915997cd70121c3d1f Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.040256 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-2cac3ab830100ddbd470ba8e091f430a3570add22a8f1f3e49a9039062f53984 WatchSource:0}: Error finding container 2cac3ab830100ddbd470ba8e091f430a3570add22a8f1f3e49a9039062f53984: Status 404 returned error can't find the container with id 2cac3ab830100ddbd470ba8e091f430a3570add22a8f1f3e49a9039062f53984 Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.046259 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.046639 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e93e253ba5b76c35126a1c6fb4316a38a1fff4cce1cf1a531b205f233b01db15 WatchSource:0}: Error finding container e93e253ba5b76c35126a1c6fb4316a38a1fff4cce1cf1a531b205f233b01db15: Status 404 returned error can't find the container with id e93e253ba5b76c35126a1c6fb4316a38a1fff4cce1cf1a531b205f233b01db15 Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.055493 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.073005 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-abdeef2f6172ccfaf8b579780c13c4444760f4a7a179e1bda4fc3e04b6ea47f7 WatchSource:0}: Error finding container abdeef2f6172ccfaf8b579780c13c4444760f4a7a179e1bda4fc3e04b6ea47f7: Status 404 returned error can't find the container with id abdeef2f6172ccfaf8b579780c13c4444760f4a7a179e1bda4fc3e04b6ea47f7 Feb 20 00:08:37 crc kubenswrapper[4750]: E0220 00:08:37.073574 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="800ms" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.080341 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-bda80c66bb67fe752d61ab7089f9c51aa3132e9fa57ee0c6c28776ee06d183f2 WatchSource:0}: Error finding container bda80c66bb67fe752d61ab7089f9c51aa3132e9fa57ee0c6c28776ee06d183f2: Status 404 returned error can't find the container with id bda80c66bb67fe752d61ab7089f9c51aa3132e9fa57ee0c6c28776ee06d183f2 Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.330142 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.332682 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.332717 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.332726 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.332748 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 00:08:37 crc kubenswrapper[4750]: E0220 00:08:37.333257 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.352331 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:37 crc kubenswrapper[4750]: E0220 00:08:37.352469 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.465828 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.466861 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 19:24:39.309129734 +0000 UTC Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.571885 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"abdeef2f6172ccfaf8b579780c13c4444760f4a7a179e1bda4fc3e04b6ea47f7"} Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.574140 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e93e253ba5b76c35126a1c6fb4316a38a1fff4cce1cf1a531b205f233b01db15"} Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.575553 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2cac3ab830100ddbd470ba8e091f430a3570add22a8f1f3e49a9039062f53984"} Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.576402 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5ca0e5998df1ce6cbde43b1a73267f375e787857bdd8a0915997cd70121c3d1f"} Feb 20 00:08:37 crc kubenswrapper[4750]: I0220 00:08:37.577044 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bda80c66bb67fe752d61ab7089f9c51aa3132e9fa57ee0c6c28776ee06d183f2"} Feb 20 00:08:37 crc kubenswrapper[4750]: E0220 00:08:37.874895 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="1.6s" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.922177 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:37 crc kubenswrapper[4750]: E0220 00:08:37.922279 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.935037 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:37 crc kubenswrapper[4750]: E0220 00:08:37.935112 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:37 crc kubenswrapper[4750]: W0220 00:08:37.998513 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:37 crc kubenswrapper[4750]: E0220 00:08:37.998812 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.134297 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.137883 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.137954 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.137974 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.138013 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 00:08:38 crc kubenswrapper[4750]: E0220 00:08:38.138676 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.465479 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.466958 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 19:43:12.943428395 +0000 UTC Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.522316 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 00:08:38 crc kubenswrapper[4750]: E0220 00:08:38.523281 4750 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.582398 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.582444 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.582457 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.582470 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.582555 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.583526 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.583561 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.583577 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.586067 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1" exitCode=0 Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.586160 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.586266 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.587138 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.587167 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.587179 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.588585 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.589427 4750 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b20ae65c36bd04bbb5ec0aeb8040eaa88ed4dc38fea385a4a28a8bff4f0cdd54" exitCode=0 Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.589501 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b20ae65c36bd04bbb5ec0aeb8040eaa88ed4dc38fea385a4a28a8bff4f0cdd54"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.589554 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.589915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.589954 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.589971 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.590778 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.590823 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.590836 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.592754 4750 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5c65262706f6f421427c770de7e6f6cd7ad157d387712a06d322e95fe6cdb1da" exitCode=0 Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.592825 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.592863 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5c65262706f6f421427c770de7e6f6cd7ad157d387712a06d322e95fe6cdb1da"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.597357 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.597430 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.597452 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.602379 4750 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4" exitCode=0 Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.602442 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4"} Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.602804 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.604329 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.604374 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.604579 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:38 crc kubenswrapper[4750]: I0220 00:08:38.736841 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.465902 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.467189 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 03:18:12.412440869 +0000 UTC Feb 20 00:08:39 crc kubenswrapper[4750]: W0220 00:08:39.475359 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:39 crc kubenswrapper[4750]: E0220 00:08:39.475424 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:39 crc kubenswrapper[4750]: E0220 00:08:39.475772 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="3.2s" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.610409 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.610487 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.610502 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.610515 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.613341 4750 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c5d4933778d05deae9413f4da7de2db52555d8c45ab9f615620192c28de2ed0f" exitCode=0 Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.613501 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c5d4933778d05deae9413f4da7de2db52555d8c45ab9f615620192c28de2ed0f"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.613524 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.615464 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.615492 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.615501 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.619240 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"59c50808a9e0dca09a04f9d1d5f572b572f289e49aab7805c66850e5d4aede64"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.619319 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.621415 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.621440 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.621450 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.625585 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.625604 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.626433 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.626470 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.626489 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707"} Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.626936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.626976 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.627020 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.628445 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.628496 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.628523 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:39 crc kubenswrapper[4750]: W0220 00:08:39.653779 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:39 crc kubenswrapper[4750]: E0220 00:08:39.653899 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.739862 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.741863 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.741907 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.741921 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:39 crc kubenswrapper[4750]: I0220 00:08:39.741951 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 00:08:39 crc kubenswrapper[4750]: E0220 00:08:39.742575 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.194:6443: connect: connection refused" node="crc" Feb 20 00:08:40 crc kubenswrapper[4750]: W0220 00:08:40.316385 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.194:6443: connect: connection refused Feb 20 00:08:40 crc kubenswrapper[4750]: E0220 00:08:40.316497 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.194:6443: connect: connection refused" logger="UnhandledError" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.468013 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 03:05:38.594749653 +0000 UTC Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.631926 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4"} Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.632188 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.632989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.633078 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.633161 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.634601 4750 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="19897b74387bef325eca11d116e254f0dfd9ab7590052bc85c0a33d4813aa253" exitCode=0 Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.634757 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.634689 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"19897b74387bef325eca11d116e254f0dfd9ab7590052bc85c0a33d4813aa253"} Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.634836 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.634898 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.634947 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.635021 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.636830 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.636904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.636945 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.636959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.636918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.637084 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.637189 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.637242 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.637263 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.637093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.637462 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:40 crc kubenswrapper[4750]: I0220 00:08:40.637481 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.036068 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.229457 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.468186 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 21:29:10.125212258 +0000 UTC Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.645341 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8aa672d6cd8364c8f4b3bfd682cb2ccf58ce2dcf0ca40d5ebc7f2de55fdba939"} Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.645416 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb5b7bfc42bb5e18149885acabe29f3c584cad8a2f2051f775a614393d3daa57"} Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.645443 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"964537c1d3bf10ce9c190ca1d5a2562025101a9fd64108658729994fadf286e5"} Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.645449 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.646737 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.646783 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:41 crc kubenswrapper[4750]: I0220 00:08:41.646800 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.468669 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 13:02:48.937955331 +0000 UTC Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.532929 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.653889 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a544220df9af054b2a1316e6ac95e1ee2b6392dd2b1b1253ed0e75d9ae256fef"} Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.653971 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ec0e5d0509cddf9e4f94f6fa2410d143f7e3a87d6a5007a02a30518300acffdc"} Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.654046 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.654049 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.655507 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.655557 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.655589 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.655633 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.655688 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.655712 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.943660 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.949922 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.949980 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.949993 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:42 crc kubenswrapper[4750]: I0220 00:08:42.950032 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.373110 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.469561 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 06:58:45.167105148 +0000 UTC Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.656243 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.657806 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.657866 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.657886 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.878788 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.879006 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.880660 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.880735 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.880754 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:43 crc kubenswrapper[4750]: I0220 00:08:43.890558 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.470163 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 21:07:28.807019664 +0000 UTC Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.658766 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.658885 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.660249 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.660300 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.660318 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.660527 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.660575 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.660592 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.851167 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.852403 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.852605 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.859239 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.859306 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:44 crc kubenswrapper[4750]: I0220 00:08:44.859333 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.470421 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 02:57:54.209473092 +0000 UTC Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.638943 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.662256 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.662256 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.664236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.664313 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.664338 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.664401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.664493 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:45 crc kubenswrapper[4750]: I0220 00:08:45.664515 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:46 crc kubenswrapper[4750]: I0220 00:08:46.471200 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 21:30:17.035363014 +0000 UTC Feb 20 00:08:46 crc kubenswrapper[4750]: E0220 00:08:46.639446 4750 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 00:08:46 crc kubenswrapper[4750]: I0220 00:08:46.914254 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:08:46 crc kubenswrapper[4750]: I0220 00:08:46.914521 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:46 crc kubenswrapper[4750]: I0220 00:08:46.916097 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:46 crc kubenswrapper[4750]: I0220 00:08:46.916198 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:46 crc kubenswrapper[4750]: I0220 00:08:46.916220 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:47 crc kubenswrapper[4750]: I0220 00:08:47.472106 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 03:06:23.630619694 +0000 UTC Feb 20 00:08:47 crc kubenswrapper[4750]: I0220 00:08:47.688243 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:47 crc kubenswrapper[4750]: I0220 00:08:47.688471 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:47 crc kubenswrapper[4750]: I0220 00:08:47.690431 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:47 crc kubenswrapper[4750]: I0220 00:08:47.690529 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:47 crc kubenswrapper[4750]: I0220 00:08:47.690553 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:47 crc kubenswrapper[4750]: I0220 00:08:47.692594 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:48 crc kubenswrapper[4750]: I0220 00:08:48.472887 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 10:57:35.600737892 +0000 UTC Feb 20 00:08:48 crc kubenswrapper[4750]: I0220 00:08:48.669802 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:48 crc kubenswrapper[4750]: I0220 00:08:48.674056 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:48 crc kubenswrapper[4750]: I0220 00:08:48.674232 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:48 crc kubenswrapper[4750]: I0220 00:08:48.674273 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:49 crc kubenswrapper[4750]: I0220 00:08:49.474095 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 20:54:20.681632379 +0000 UTC Feb 20 00:08:50 crc kubenswrapper[4750]: I0220 00:08:50.466827 4750 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 20 00:08:50 crc kubenswrapper[4750]: I0220 00:08:50.474530 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 14:13:24.272798953 +0000 UTC Feb 20 00:08:50 crc kubenswrapper[4750]: I0220 00:08:50.688339 4750 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 00:08:50 crc kubenswrapper[4750]: I0220 00:08:50.688428 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 00:08:50 crc kubenswrapper[4750]: W0220 00:08:50.892632 4750 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 20 00:08:50 crc kubenswrapper[4750]: I0220 00:08:50.892766 4750 trace.go:236] Trace[43756229]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 00:08:40.890) (total time: 10001ms): Feb 20 00:08:50 crc kubenswrapper[4750]: Trace[43756229]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:08:50.892) Feb 20 00:08:50 crc kubenswrapper[4750]: Trace[43756229]: [10.001898096s] [10.001898096s] END Feb 20 00:08:50 crc kubenswrapper[4750]: E0220 00:08:50.892806 4750 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 20 00:08:51 crc kubenswrapper[4750]: I0220 00:08:51.331202 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 20 00:08:51 crc kubenswrapper[4750]: I0220 00:08:51.331272 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 00:08:51 crc kubenswrapper[4750]: I0220 00:08:51.344237 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 20 00:08:51 crc kubenswrapper[4750]: I0220 00:08:51.344351 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 00:08:51 crc kubenswrapper[4750]: I0220 00:08:51.475288 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 00:19:19.88306259 +0000 UTC Feb 20 00:08:52 crc kubenswrapper[4750]: I0220 00:08:52.476411 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 22:41:41.911654382 +0000 UTC Feb 20 00:08:53 crc kubenswrapper[4750]: I0220 00:08:53.476590 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 20:06:02.680988491 +0000 UTC Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.477566 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 04:22:43.984615129 +0000 UTC Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.861534 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.861815 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.863783 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.863849 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.863873 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.869449 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.893261 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.893545 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.895262 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.895325 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.895347 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:54 crc kubenswrapper[4750]: I0220 00:08:54.915411 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.478142 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 05:38:33.773819885 +0000 UTC Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.689160 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.689277 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.689177 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.691236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.691288 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.691309 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.692165 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.692281 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:08:55 crc kubenswrapper[4750]: I0220 00:08:55.692308 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.328820 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.331281 4750 trace.go:236] Trace[1069950922]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 00:08:44.818) (total time: 11512ms): Feb 20 00:08:56 crc kubenswrapper[4750]: Trace[1069950922]: ---"Objects listed" error: 11512ms (00:08:56.331) Feb 20 00:08:56 crc kubenswrapper[4750]: Trace[1069950922]: [11.512347964s] [11.512347964s] END Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.331322 4750 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.333490 4750 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.343570 4750 trace.go:236] Trace[1723019659]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 00:08:46.282) (total time: 10061ms): Feb 20 00:08:56 crc kubenswrapper[4750]: Trace[1723019659]: ---"Objects listed" error: 10061ms (00:08:56.343) Feb 20 00:08:56 crc kubenswrapper[4750]: Trace[1723019659]: [10.061442875s] [10.061442875s] END Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.343630 4750 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.343596 4750 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.344536 4750 trace.go:236] Trace[1531793592]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 00:08:42.736) (total time: 13607ms): Feb 20 00:08:56 crc kubenswrapper[4750]: Trace[1531793592]: ---"Objects listed" error: 13607ms (00:08:56.343) Feb 20 00:08:56 crc kubenswrapper[4750]: Trace[1531793592]: [13.607557589s] [13.607557589s] END Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.344591 4750 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.352804 4750 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.375219 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36876->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.375308 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36876->192.168.126.11:17697: read: connection reset by peer" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.375664 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.375736 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.455748 4750 apiserver.go:52] "Watching apiserver" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.463647 4750 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.463951 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.464425 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.464482 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.464258 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.466403 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.466519 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.466645 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.466752 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.468267 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.468666 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.473360 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.473876 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.473887 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.474085 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.474183 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.474227 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.474332 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.478592 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 07:46:59.799287581 +0000 UTC Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.479153 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.479192 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.512398 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.519380 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.519446 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.524734 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.538323 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.544799 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.545045 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.545264 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.545408 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.545544 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:57.045486976 +0000 UTC m=+21.240323035 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.545432 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.545973 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.546264 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.546599 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.546784 4750 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.547462 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.547549 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.547924 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.547954 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.548062 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.548084 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.548099 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.548130 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.548035 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.548329 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:57.048314231 +0000 UTC m=+21.243150280 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.548488 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.549095 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.553466 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.556444 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.557380 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.564411 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.564450 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.564466 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.564549 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:57.064524952 +0000 UTC m=+21.259361011 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.568520 4750 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.568621 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.568674 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.568698 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.568786 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:57.068754174 +0000 UTC m=+21.263590263 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.574433 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.574683 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.579764 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.580302 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.592397 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.607469 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.619455 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.631177 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.648876 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.648938 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649043 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649156 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649211 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649289 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649298 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649298 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649578 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.649588 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.650088 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.650165 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.650283 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.650706 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651044 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651634 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651693 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651727 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651759 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651790 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651823 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651860 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651896 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651930 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651964 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.651998 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652027 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652061 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652064 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652092 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652152 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652189 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652224 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652260 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652293 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652326 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652361 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652400 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652442 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652474 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652506 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652540 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652576 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652575 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652630 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652662 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652696 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652735 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652774 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652811 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652827 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652853 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652888 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652926 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652959 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.652991 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653035 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653071 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653165 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653226 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653254 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653263 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653299 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653332 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653387 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653423 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653455 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653491 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653537 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653571 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653604 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653638 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653673 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653696 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653707 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653739 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653776 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653811 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653845 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653875 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653909 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653931 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653947 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653971 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.653996 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654021 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654045 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654069 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654093 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654144 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654167 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654193 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654221 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654244 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654267 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654311 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654333 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654360 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654394 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654454 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654487 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654542 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654576 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654608 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654641 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654672 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654706 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654739 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654772 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654805 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654836 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654843 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654870 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654904 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654938 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.654974 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655008 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655089 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655154 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655192 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655230 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655266 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655301 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655337 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655374 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655413 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655449 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655485 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655519 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655554 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655587 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655627 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655661 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655697 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655734 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655772 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655806 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655842 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655875 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655908 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655942 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655974 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656005 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656039 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656074 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656109 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656169 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656206 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656241 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656275 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656310 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656345 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656381 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656417 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656452 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656492 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656533 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656571 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656608 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656642 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656677 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656722 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656760 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656802 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656839 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656882 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656920 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656956 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656993 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657026 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657062 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657103 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657164 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657210 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657358 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657399 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657436 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657470 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657513 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657552 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657586 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657621 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657655 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657691 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657724 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657758 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657794 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657835 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657899 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657940 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657976 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658013 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658053 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658088 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658149 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658176 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658201 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658226 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658249 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659009 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659219 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659321 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659417 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659682 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659835 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661163 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661219 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661252 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661315 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661347 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661478 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662152 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662315 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662350 4750 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662370 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662389 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662409 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662429 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662450 4750 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662471 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662492 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662544 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662569 4750 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662592 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662612 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662629 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.665672 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.674219 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.675424 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655090 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655365 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655617 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.655877 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656331 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.656959 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657342 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657621 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.676609 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.657915 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658180 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.658839 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659291 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659438 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659425 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659643 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.659933 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.660078 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.660300 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.660463 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.660753 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.660782 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.660931 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661361 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661141 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661790 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.661838 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662162 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662232 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662531 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.662845 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.663043 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.663337 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.663501 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.663593 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.663609 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.663907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.663949 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.664229 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.665417 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.666188 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.666529 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.666586 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.666613 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.666827 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.666888 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: E0220 00:08:56.666970 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:08:57.166945411 +0000 UTC m=+21.361781470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.666959 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.667071 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.667960 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.668349 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.668510 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.668701 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.668823 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.669103 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.669844 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.670266 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.670700 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671004 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671059 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671062 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671403 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671443 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671677 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671719 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671751 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.671783 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.672106 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.672302 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.672331 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.672571 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.673264 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.673352 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.673390 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.673778 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.673802 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.674145 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.674873 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.675401 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.675988 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.676160 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.676319 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.676518 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.676884 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.676945 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.676974 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.677263 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.677429 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.677652 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.677699 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.678307 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.678461 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.678704 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.679369 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.679907 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680018 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680045 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680189 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680515 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680705 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680856 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680871 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.680982 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.681595 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.681921 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.681949 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.681992 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.682047 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.682078 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.682103 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.682200 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.682998 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.683263 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.683290 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.683303 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.683509 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.683682 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.683852 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.684005 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.684052 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.684267 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.684871 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.685000 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.685416 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.685905 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.686335 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.686455 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.687411 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.687498 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.687684 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.687823 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.687853 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.688012 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.688233 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.688318 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.688515 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.688914 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689164 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689269 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689905 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689947 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689625 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689652 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689698 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689709 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689868 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.690051 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.690346 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.690617 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.690745 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.690851 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.689701 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.693216 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.693634 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.694133 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.694241 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.694409 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.695550 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.695633 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.696723 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.696810 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.696710 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.697080 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.697246 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.697519 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.697857 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.697965 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.698170 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.698248 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.698436 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.698461 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.699097 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.699484 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.699538 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.700345 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.700361 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.700368 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.700601 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.701354 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.701490 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.701804 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4" exitCode=255 Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.701869 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4"} Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.713806 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.718758 4750 scope.go:117] "RemoveContainer" containerID="5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.720900 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.726817 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.737076 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.743408 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.750743 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.750805 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.753875 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.762297 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764169 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764218 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764243 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764265 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764287 4750 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764309 4750 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764329 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764350 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764370 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764391 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764411 4750 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764434 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764454 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764474 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764495 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764515 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764536 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764556 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764576 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764595 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764617 4750 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764639 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764660 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764680 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764700 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764720 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764740 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764764 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764785 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764809 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764828 4750 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764848 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764869 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764889 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764922 4750 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764943 4750 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764964 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.764986 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765004 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765023 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765042 4750 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765061 4750 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765081 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765101 4750 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765146 4750 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765167 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765189 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765209 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765230 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765250 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765272 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765291 4750 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765309 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765329 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765350 4750 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765372 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765391 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765411 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765432 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765485 4750 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765508 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765528 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765547 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765566 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765586 4750 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765606 4750 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765625 4750 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765643 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765663 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765680 4750 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765699 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765721 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765741 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765761 4750 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765781 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765802 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765822 4750 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765842 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765862 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765882 4750 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765902 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765923 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765942 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765961 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.765980 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766000 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766020 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766038 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766057 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766077 4750 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766096 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766140 4750 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766162 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766183 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766207 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766234 4750 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766260 4750 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766287 4750 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766313 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766339 4750 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766367 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766389 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766408 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766428 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766448 4750 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766467 4750 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766488 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766549 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766579 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766607 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766636 4750 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766663 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766691 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766716 4750 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766740 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766766 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766794 4750 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766815 4750 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766834 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766856 4750 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766875 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766898 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766917 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766935 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766965 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.766984 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767004 4750 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767023 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767041 4750 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767059 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767077 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767096 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767145 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767165 4750 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767182 4750 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767200 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767221 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767239 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767258 4750 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767276 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767294 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767312 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767330 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767348 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767372 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767396 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767423 4750 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767441 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767460 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767480 4750 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767499 4750 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767517 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767536 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767558 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767581 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767607 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767628 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767648 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767666 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767684 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767703 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767720 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767741 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767759 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767776 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767796 4750 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767813 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767830 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767848 4750 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767864 4750 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767882 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767902 4750 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767926 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767967 4750 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.767989 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.768007 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.768025 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.768043 4750 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.768061 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.768079 4750 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.781225 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.790154 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.792288 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.804846 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 00:08:56 crc kubenswrapper[4750]: I0220 00:08:56.817273 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.051632 4750 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.071425 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.071491 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.071535 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.071572 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.071745 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.071798 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.071828 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.071898 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:58.071875513 +0000 UTC m=+22.266711602 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072043 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072152 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072171 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072257 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:58.072230253 +0000 UTC m=+22.267066302 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072322 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072362 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:58.072353586 +0000 UTC m=+22.267189635 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072462 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.072608 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:08:58.072583642 +0000 UTC m=+22.267419861 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.172526 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.172714 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:08:58.17268464 +0000 UTC m=+22.367520689 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.479524 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 22:07:41.412394291 +0000 UTC Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.558898 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:57 crc kubenswrapper[4750]: E0220 00:08:57.559083 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.696952 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.703898 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.707727 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928"} Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.707787 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"22565a20c21c803898af6b6503fcb28a09f7ae1d26d4a096338503c5bf5deecc"} Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.713548 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.714745 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.720838 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd"} Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.721152 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.721435 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.722198 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"991ee74bcee5ebf78e4e654ddc2df7ad43658b4fc1345c5425bcdefb9923f533"} Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.724278 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46"} Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.724337 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b"} Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.724365 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e316c0f264c1e67695a3d6f58f9eade8fb0ac850e619b2d589f6c33051e3f170"} Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.737020 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.756492 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.778160 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.799228 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.817574 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.833738 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.852281 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.868182 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.879865 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.893289 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.902740 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.913844 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.926268 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:57 crc kubenswrapper[4750]: I0220 00:08:57.937779 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.079942 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.080012 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.080052 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.080103 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080150 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080193 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080203 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080218 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080265 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080277 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080288 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080277 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080253 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:00.080229278 +0000 UTC m=+24.275065357 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080338 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:00.080327 +0000 UTC m=+24.275163049 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080352 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:00.080345341 +0000 UTC m=+24.275181390 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.080364 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:00.080358141 +0000 UTC m=+24.275194190 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.181311 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.181484 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:09:00.181466156 +0000 UTC m=+24.376302225 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.479898 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 22:47:11.118525717 +0000 UTC Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.559424 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.559501 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.559617 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:08:58 crc kubenswrapper[4750]: E0220 00:08:58.559682 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.566958 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.568559 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.571695 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.573057 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.575088 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.576244 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.577436 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.579545 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.581237 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.583482 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.584524 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.586720 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.587859 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.589030 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.591187 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.592285 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.594812 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.595662 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.596912 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.599022 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.600107 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.602556 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.604203 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.606699 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.607372 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.608172 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.609582 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.610284 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.611643 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.612258 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.613379 4750 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.613514 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.615797 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.617400 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.617949 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.619965 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.620815 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.622264 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.623429 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.624872 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.625495 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.626813 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.627738 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.629020 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.629647 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.630875 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.631584 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.633008 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.633628 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.634728 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.635465 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.636646 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.637455 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 20 00:08:58 crc kubenswrapper[4750]: I0220 00:08:58.638057 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 20 00:08:59 crc kubenswrapper[4750]: I0220 00:08:59.481060 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 18:25:16.618376444 +0000 UTC Feb 20 00:08:59 crc kubenswrapper[4750]: I0220 00:08:59.559780 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:08:59 crc kubenswrapper[4750]: E0220 00:08:59.559920 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.097863 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.097920 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.097956 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.097986 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098098 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098170 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098214 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098272 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098292 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098226 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098370 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098235 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:04.09820955 +0000 UTC m=+28.293045609 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098450 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:04.098401005 +0000 UTC m=+28.293237214 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098473 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:04.098463307 +0000 UTC m=+28.293299556 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098379 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.098523 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:04.098513498 +0000 UTC m=+28.293349847 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.198520 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.198744 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:09:04.198687438 +0000 UTC m=+28.393523507 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.481630 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 10:49:40.531616178 +0000 UTC Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.559701 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.559941 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.560105 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:00 crc kubenswrapper[4750]: E0220 00:09:00.560874 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.734462 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826"} Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.759706 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.783437 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.807436 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.825453 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.839830 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.853639 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.869802 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:00 crc kubenswrapper[4750]: I0220 00:09:00.888788 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:00Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:01 crc kubenswrapper[4750]: I0220 00:09:01.482341 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 03:22:34.296797774 +0000 UTC Feb 20 00:09:01 crc kubenswrapper[4750]: I0220 00:09:01.559152 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:01 crc kubenswrapper[4750]: E0220 00:09:01.559307 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:01 crc kubenswrapper[4750]: I0220 00:09:01.594727 4750 csr.go:261] certificate signing request csr-xtfj9 is approved, waiting to be issued Feb 20 00:09:01 crc kubenswrapper[4750]: I0220 00:09:01.622383 4750 csr.go:257] certificate signing request csr-xtfj9 is issued Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.121680 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-dm8lb"] Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.122013 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-fk2zg"] Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.122158 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.122251 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xjx4s"] Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.122579 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.122974 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.126144 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.126633 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.126862 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.126990 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.127135 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hkqgt"] Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.127143 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.127361 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.127308 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.127585 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.127656 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.130091 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.131379 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.136214 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.138468 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-96tjn"] Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.138688 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.138768 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.141179 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.142868 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.142873 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.142913 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.143567 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.143631 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.143681 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.143638 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.145173 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.145270 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.149808 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.170392 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.183736 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.195896 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.210530 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216050 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-netd\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216094 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-k8s-cni-cncf-io\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216137 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-os-release\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216160 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-cni-multus\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216194 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-kubelet\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216244 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-systemd\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216296 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62d0755e-ca22-4187-aae7-65377cedcaa9-proxy-tls\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216353 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-cni-bin\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216381 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216403 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-config\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216427 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216454 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02349b76-be13-4b84-a010-38792f5d9d86-cni-binary-copy\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216474 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-netns\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216493 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-hostroot\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216513 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-slash\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216531 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-var-lib-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216551 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-node-log\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216572 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-kubelet\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216603 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62d0755e-ca22-4187-aae7-65377cedcaa9-mcd-auth-proxy-config\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216626 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-etc-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216656 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-daemon-config\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216676 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-env-overrides\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216696 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5lm\" (UniqueName: \"kubernetes.io/projected/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-kube-api-access-xm5lm\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216717 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlzf8\" (UniqueName: \"kubernetes.io/projected/18b1268a-b190-41c7-9746-a801bf44bd8b-kube-api-access-qlzf8\") pod \"node-resolver-96tjn\" (UID: \"18b1268a-b190-41c7-9746-a801bf44bd8b\") " pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216739 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-os-release\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216759 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-cni-binary-copy\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216781 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-etc-kubernetes\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216801 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/62d0755e-ca22-4187-aae7-65377cedcaa9-rootfs\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216823 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-socket-dir-parent\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216846 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-system-cni-dir\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216867 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216888 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-cni-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216907 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-log-socket\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216929 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovn-node-metrics-cert\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216948 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-script-lib\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216970 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-ovn\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.216993 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02349b76-be13-4b84-a010-38792f5d9d86-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217012 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-multus-certs\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217033 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tdtm\" (UniqueName: \"kubernetes.io/projected/62d0755e-ca22-4187-aae7-65377cedcaa9-kube-api-access-8tdtm\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217052 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-systemd-units\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217074 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-ovn-kubernetes\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217094 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-bin\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217142 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-cnibin\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217164 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-conf-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217186 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzxkk\" (UniqueName: \"kubernetes.io/projected/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-kube-api-access-lzxkk\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217206 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-netns\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217227 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-cnibin\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217257 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-system-cni-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217276 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/18b1268a-b190-41c7-9746-a801bf44bd8b-hosts-file\") pod \"node-resolver-96tjn\" (UID: \"18b1268a-b190-41c7-9746-a801bf44bd8b\") " pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.217306 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfpmg\" (UniqueName: \"kubernetes.io/projected/02349b76-be13-4b84-a010-38792f5d9d86-kube-api-access-cfpmg\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.228159 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.247853 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.263787 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.275268 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.288858 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.301490 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.314605 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317723 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-node-log\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317759 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-kubelet\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317794 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-var-lib-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317824 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62d0755e-ca22-4187-aae7-65377cedcaa9-mcd-auth-proxy-config\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317835 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-node-log\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317846 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-etc-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317871 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-etc-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.317906 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-kubelet\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318015 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-var-lib-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318105 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-daemon-config\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318161 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-env-overrides\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318185 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlzf8\" (UniqueName: \"kubernetes.io/projected/18b1268a-b190-41c7-9746-a801bf44bd8b-kube-api-access-qlzf8\") pod \"node-resolver-96tjn\" (UID: \"18b1268a-b190-41c7-9746-a801bf44bd8b\") " pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318205 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-os-release\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318221 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-cni-binary-copy\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318237 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-etc-kubernetes\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318253 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/62d0755e-ca22-4187-aae7-65377cedcaa9-rootfs\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318270 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5lm\" (UniqueName: \"kubernetes.io/projected/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-kube-api-access-xm5lm\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318307 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-system-cni-dir\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318328 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318344 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-socket-dir-parent\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318362 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-cni-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318385 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-log-socket\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318402 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovn-node-metrics-cert\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318417 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-script-lib\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318435 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02349b76-be13-4b84-a010-38792f5d9d86-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318457 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-multus-certs\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318482 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tdtm\" (UniqueName: \"kubernetes.io/projected/62d0755e-ca22-4187-aae7-65377cedcaa9-kube-api-access-8tdtm\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318499 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-systemd-units\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318514 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-ovn\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318533 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-bin\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318557 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-cnibin\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318573 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-conf-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318589 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzxkk\" (UniqueName: \"kubernetes.io/projected/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-kube-api-access-lzxkk\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318604 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-netns\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318621 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-ovn-kubernetes\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318639 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-cnibin\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318653 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-system-cni-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318672 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/18b1268a-b190-41c7-9746-a801bf44bd8b-hosts-file\") pod \"node-resolver-96tjn\" (UID: \"18b1268a-b190-41c7-9746-a801bf44bd8b\") " pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318689 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfpmg\" (UniqueName: \"kubernetes.io/projected/02349b76-be13-4b84-a010-38792f5d9d86-kube-api-access-cfpmg\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318708 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-k8s-cni-cncf-io\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318724 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-netd\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318740 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-os-release\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318756 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-cni-multus\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-kubelet\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318785 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-systemd\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318786 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-os-release\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318802 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62d0755e-ca22-4187-aae7-65377cedcaa9-proxy-tls\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318833 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-daemon-config\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318853 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-cni-bin\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318905 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-netns\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318941 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-ovn-kubernetes\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318950 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-config\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318961 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-cni-binary-copy\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318979 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02349b76-be13-4b84-a010-38792f5d9d86-cni-binary-copy\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318987 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-cnibin\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319017 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-cni-bin\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319029 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-system-cni-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319016 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-netns\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319042 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-netns\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319061 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-hostroot\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319087 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319129 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-slash\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319132 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-netd\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319182 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-os-release\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318883 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-env-overrides\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319212 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319066 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/18b1268a-b190-41c7-9746-a801bf44bd8b-hosts-file\") pod \"node-resolver-96tjn\" (UID: \"18b1268a-b190-41c7-9746-a801bf44bd8b\") " pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319238 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-hostroot\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.318922 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-k8s-cni-cncf-io\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319260 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-openvswitch\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319088 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-cni-multus\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319304 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-slash\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319337 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-ovn\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319367 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-cnibin\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319390 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-script-lib\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319400 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-bin\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319433 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-conf-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319434 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-system-cni-dir\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319450 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-etc-kubernetes\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319475 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/62d0755e-ca22-4187-aae7-65377cedcaa9-rootfs\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319689 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-run-multus-certs\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319719 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-systemd-units\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319827 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-config\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319859 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-socket-dir-parent\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319890 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-host-var-lib-kubelet\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319895 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/02349b76-be13-4b84-a010-38792f5d9d86-cni-binary-copy\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319925 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-systemd\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.319949 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-log-socket\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.320004 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-multus-cni-dir\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.320263 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02349b76-be13-4b84-a010-38792f5d9d86-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.320379 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62d0755e-ca22-4187-aae7-65377cedcaa9-mcd-auth-proxy-config\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.322341 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovn-node-metrics-cert\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.324586 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02349b76-be13-4b84-a010-38792f5d9d86-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.327682 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62d0755e-ca22-4187-aae7-65377cedcaa9-proxy-tls\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.342176 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlzf8\" (UniqueName: \"kubernetes.io/projected/18b1268a-b190-41c7-9746-a801bf44bd8b-kube-api-access-qlzf8\") pod \"node-resolver-96tjn\" (UID: \"18b1268a-b190-41c7-9746-a801bf44bd8b\") " pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.344427 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzxkk\" (UniqueName: \"kubernetes.io/projected/1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030-kube-api-access-lzxkk\") pod \"multus-dm8lb\" (UID: \"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\") " pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.345484 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5lm\" (UniqueName: \"kubernetes.io/projected/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-kube-api-access-xm5lm\") pod \"ovnkube-node-hkqgt\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.346170 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.360627 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tdtm\" (UniqueName: \"kubernetes.io/projected/62d0755e-ca22-4187-aae7-65377cedcaa9-kube-api-access-8tdtm\") pod \"machine-config-daemon-fk2zg\" (UID: \"62d0755e-ca22-4187-aae7-65377cedcaa9\") " pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.365713 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfpmg\" (UniqueName: \"kubernetes.io/projected/02349b76-be13-4b84-a010-38792f5d9d86-kube-api-access-cfpmg\") pod \"multus-additional-cni-plugins-xjx4s\" (UID: \"02349b76-be13-4b84-a010-38792f5d9d86\") " pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.371692 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.386193 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.423947 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.440670 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dm8lb" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.446417 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.449936 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:09:02 crc kubenswrapper[4750]: W0220 00:09:02.453944 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1acbd435_6fc9_40bf_b2b4_1f8ff1b9c030.slice/crio-75fead914ba257a8dbf37780b8bf9b402ebfb58e8b1e3ab7e86e8f6f4804e44a WatchSource:0}: Error finding container 75fead914ba257a8dbf37780b8bf9b402ebfb58e8b1e3ab7e86e8f6f4804e44a: Status 404 returned error can't find the container with id 75fead914ba257a8dbf37780b8bf9b402ebfb58e8b1e3ab7e86e8f6f4804e44a Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.458298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.466354 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.474298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-96tjn" Feb 20 00:09:02 crc kubenswrapper[4750]: W0220 00:09:02.476678 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02349b76_be13_4b84_a010_38792f5d9d86.slice/crio-a43fc9443ec43f6350372035bae596d2ec62179f6ccbf981e1bd45ba612e038d WatchSource:0}: Error finding container a43fc9443ec43f6350372035bae596d2ec62179f6ccbf981e1bd45ba612e038d: Status 404 returned error can't find the container with id a43fc9443ec43f6350372035bae596d2ec62179f6ccbf981e1bd45ba612e038d Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.479040 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.483192 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 15:32:17.187886661 +0000 UTC Feb 20 00:09:02 crc kubenswrapper[4750]: W0220 00:09:02.503613 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17ae630a_7ade_407c_b93c_e4bae6d0f8e3.slice/crio-22ef73bc6fd4ef403a25b7711e8495189441418f3c270591532b3048b06e9e63 WatchSource:0}: Error finding container 22ef73bc6fd4ef403a25b7711e8495189441418f3c270591532b3048b06e9e63: Status 404 returned error can't find the container with id 22ef73bc6fd4ef403a25b7711e8495189441418f3c270591532b3048b06e9e63 Feb 20 00:09:02 crc kubenswrapper[4750]: W0220 00:09:02.506315 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18b1268a_b190_41c7_9746_a801bf44bd8b.slice/crio-deb31466aa7b86dd463ebdd3d55ad36ddf369a21256db1d98e9aca147c711b0e WatchSource:0}: Error finding container deb31466aa7b86dd463ebdd3d55ad36ddf369a21256db1d98e9aca147c711b0e: Status 404 returned error can't find the container with id deb31466aa7b86dd463ebdd3d55ad36ddf369a21256db1d98e9aca147c711b0e Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.514051 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.531486 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.553970 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.560223 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.560316 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.560721 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.560823 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.583534 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.624033 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-20 00:04:01 +0000 UTC, rotation deadline is 2026-12-19 05:01:43.074300392 +0000 UTC Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.624109 4750 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7252h52m40.450194413s for next certificate rotation Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.734546 4750 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.738636 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.738664 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.738672 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.738765 4750 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.741811 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerStarted","Data":"dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.741876 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerStarted","Data":"a43fc9443ec43f6350372035bae596d2ec62179f6ccbf981e1bd45ba612e038d"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.743410 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerStarted","Data":"bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.743462 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerStarted","Data":"75fead914ba257a8dbf37780b8bf9b402ebfb58e8b1e3ab7e86e8f6f4804e44a"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.745079 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.745143 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.745160 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"98aacd446e780a3ad7b9b3247cd32c86252c3b6b1f04e06968067c0e3c0bc791"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.746552 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-96tjn" event={"ID":"18b1268a-b190-41c7-9746-a801bf44bd8b","Type":"ContainerStarted","Data":"11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.746596 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-96tjn" event={"ID":"18b1268a-b190-41c7-9746-a801bf44bd8b","Type":"ContainerStarted","Data":"deb31466aa7b86dd463ebdd3d55ad36ddf369a21256db1d98e9aca147c711b0e"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.747804 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5" exitCode=0 Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.747848 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.747875 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"22ef73bc6fd4ef403a25b7711e8495189441418f3c270591532b3048b06e9e63"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.758077 4750 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.758396 4750 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.759658 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.759702 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.759712 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.759732 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.759743 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:02Z","lastTransitionTime":"2026-02-20T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.763602 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.777766 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.780574 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.786784 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.786845 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.786861 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.786888 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.786940 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:02Z","lastTransitionTime":"2026-02-20T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.805726 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.809767 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.809813 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.809822 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.809838 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.809849 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:02Z","lastTransitionTime":"2026-02-20T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.814361 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.823515 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.827012 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.827051 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.827060 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.827083 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.827102 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:02Z","lastTransitionTime":"2026-02-20T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.835693 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.843455 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.846800 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.846833 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.846845 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.846864 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.846877 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:02Z","lastTransitionTime":"2026-02-20T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.849085 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.860853 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: E0220 00:09:02.860999 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.862824 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.862866 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.862876 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.862895 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.862924 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:02Z","lastTransitionTime":"2026-02-20T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.862870 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.876416 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.888710 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.904336 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.914383 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.925854 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.937456 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.955653 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.969523 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.969564 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.969574 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.969592 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.969604 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:02Z","lastTransitionTime":"2026-02-20T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.973101 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:02 crc kubenswrapper[4750]: I0220 00:09:02.990657 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:02Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.011296 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.026131 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.040970 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.052281 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.061825 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.074632 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.074704 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.074718 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.074740 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.074757 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.074620 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.097128 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.109237 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.122523 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.139719 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.151510 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.178312 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.178361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.178372 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.178397 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.178410 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.281231 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.281271 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.281283 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.281308 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.281339 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.383619 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.383646 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.383654 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.383668 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.383677 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.484326 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 20:40:11.604241924 +0000 UTC Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.486373 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.486397 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.486407 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.486423 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.486553 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.559425 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:03 crc kubenswrapper[4750]: E0220 00:09:03.559555 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.588830 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.588862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.588871 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.588885 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.588900 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.690926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.690967 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.690999 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.691062 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.691073 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.754818 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.755394 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.755414 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.755426 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.755439 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.755474 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.756302 4750 generic.go:334] "Generic (PLEG): container finished" podID="02349b76-be13-4b84-a010-38792f5d9d86" containerID="dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7" exitCode=0 Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.756330 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerDied","Data":"dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.780859 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.793741 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.793779 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.793788 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.793803 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.793814 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.798729 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.811687 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.827313 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.839601 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.853326 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.875185 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.887657 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.897545 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.897590 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.897602 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.897620 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.897632 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:03Z","lastTransitionTime":"2026-02-20T00:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.899934 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.911913 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.926162 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.940067 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:03 crc kubenswrapper[4750]: I0220 00:09:03.950597 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:03Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.001967 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.002002 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.002012 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.002027 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.002038 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.105232 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.105281 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.105295 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.105313 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.105325 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.139391 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.139467 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.139518 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.139568 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139605 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139651 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139673 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139723 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139750 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:12.139723523 +0000 UTC m=+36.334559612 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139802 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:12.139777364 +0000 UTC m=+36.334613453 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139900 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139920 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139940 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.139988 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:12.139973089 +0000 UTC m=+36.334809178 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.140046 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.140089 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:12.140075522 +0000 UTC m=+36.334911621 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.208510 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.208819 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.208957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.209106 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.209262 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.240465 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.240694 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:09:12.240654823 +0000 UTC m=+36.435490912 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.311557 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.311617 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.311634 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.311660 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.311677 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.414433 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.414503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.414523 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.414550 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.414568 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.484996 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 05:19:38.506692052 +0000 UTC Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.517821 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.519343 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.519503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.519653 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.519791 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.559263 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.559298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.559404 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:04 crc kubenswrapper[4750]: E0220 00:09:04.559484 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.622052 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.622449 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.622470 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.622494 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.622511 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.724662 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.724837 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.724904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.724973 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.725041 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.761987 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerStarted","Data":"e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.775361 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.789606 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.814220 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.827874 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.828094 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.828206 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.828302 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.828386 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.832811 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.851799 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.862931 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.879588 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.899767 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.918552 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.930860 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.930905 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.930918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.930936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.930949 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:04Z","lastTransitionTime":"2026-02-20T00:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.931402 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.948063 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.961156 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:04 crc kubenswrapper[4750]: I0220 00:09:04.972712 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.033888 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.033937 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.033948 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.033967 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.033977 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.136962 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.136997 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.137006 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.137019 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.137028 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.200100 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-q46qb"] Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.200813 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.203780 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.204143 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.204412 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.204519 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.221589 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.238645 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.240231 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.240288 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.240305 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.240327 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.240342 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.252432 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd97j\" (UniqueName: \"kubernetes.io/projected/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-kube-api-access-nd97j\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.252599 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-host\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.252718 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-serviceca\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.253655 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.267255 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.280606 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.295823 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.314769 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.337439 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.342824 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.342874 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.342890 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.342912 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.342926 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.353676 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-host\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.353734 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-serviceca\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.353773 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd97j\" (UniqueName: \"kubernetes.io/projected/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-kube-api-access-nd97j\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.353755 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.354048 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-host\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.355263 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-serviceca\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.369707 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.371056 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd97j\" (UniqueName: \"kubernetes.io/projected/46e36f94-f9ca-490e-9144-4d4fd2bf5eeb-kube-api-access-nd97j\") pod \"node-ca-q46qb\" (UID: \"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\") " pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.387613 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.404225 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.418940 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.430478 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.446623 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.446674 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.446686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.446704 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.446717 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.486282 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 07:14:40.436220595 +0000 UTC Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.515387 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-q46qb" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.549295 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.549332 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.549345 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.549364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.549378 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.559035 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:05 crc kubenswrapper[4750]: E0220 00:09:05.559617 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.652571 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.652611 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.652620 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.652637 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.652646 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.756212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.756266 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.756279 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.756297 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.756309 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.768180 4750 generic.go:334] "Generic (PLEG): container finished" podID="02349b76-be13-4b84-a010-38792f5d9d86" containerID="e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882" exitCode=0 Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.768272 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerDied","Data":"e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.772549 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q46qb" event={"ID":"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb","Type":"ContainerStarted","Data":"175aee82381437a91d9ad64a823e9024cb04ac15a62cb9ae61e58365e0e3ab40"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.790831 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.805913 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.826642 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.841315 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.856432 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.863431 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.863476 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.863489 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.863510 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.863524 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.870856 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.891261 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.903922 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.965964 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.965994 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.966004 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.966018 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:05 crc kubenswrapper[4750]: I0220 00:09:05.966029 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:05Z","lastTransitionTime":"2026-02-20T00:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.068171 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.068241 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.068255 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.068278 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.068293 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.171099 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.171182 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.171220 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.171248 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.171265 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.274167 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.274214 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.274229 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.274249 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.274265 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.300487 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.317996 4750 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.320758 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-operator/pods/network-operator-58b4c7f79c-55gtf/status\": read tcp 38.102.83.194:42672->38.102.83.194:6443: use of closed network connection" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.385354 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.385380 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.385487 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.385497 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.385511 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.385522 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.400177 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.414179 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.425384 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.489027 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 18:29:55.791543597 +0000 UTC Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.491315 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.491361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.491370 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.491388 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.491399 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.558986 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:06 crc kubenswrapper[4750]: E0220 00:09:06.560077 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.559048 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:06 crc kubenswrapper[4750]: E0220 00:09:06.560483 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.574891 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.593719 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.593836 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.593915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.594011 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.594453 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.595300 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.616968 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.641408 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.658397 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.674657 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.692001 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.697285 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.697321 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.697334 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.697351 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.697362 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.705910 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.720875 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.733104 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.770481 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.787375 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.789787 4750 generic.go:334] "Generic (PLEG): container finished" podID="02349b76-be13-4b84-a010-38792f5d9d86" containerID="a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332" exitCode=0 Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.789835 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerDied","Data":"a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.799308 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.799353 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.799370 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.799392 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.799409 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.800601 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.802603 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-q46qb" event={"ID":"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb","Type":"ContainerStarted","Data":"af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.805711 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.832982 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.859354 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.877417 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.892952 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.901825 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.901879 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.901896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.901917 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.901934 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:06Z","lastTransitionTime":"2026-02-20T00:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.920526 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.938583 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.953681 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.969193 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:06 crc kubenswrapper[4750]: I0220 00:09:06.990380 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.004237 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.004286 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.004297 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.004312 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.004321 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.012996 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.029452 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.050658 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.069478 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.085579 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.099930 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.106483 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.106548 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.106565 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.106590 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.106610 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.208451 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.208492 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.208501 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.208517 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.208527 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.311025 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.311061 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.311073 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.311087 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.311095 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.413751 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.413817 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.413834 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.413862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.413879 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.489975 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 12:25:25.228787286 +0000 UTC Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.517281 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.517336 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.517352 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.517377 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.517394 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.559902 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:07 crc kubenswrapper[4750]: E0220 00:09:07.560096 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.620267 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.620335 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.620354 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.620381 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.620398 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.723585 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.723667 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.723686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.723711 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.723729 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.810476 4750 generic.go:334] "Generic (PLEG): container finished" podID="02349b76-be13-4b84-a010-38792f5d9d86" containerID="2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8" exitCode=0 Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.810556 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerDied","Data":"2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.824107 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.826866 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.826899 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.826909 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.826926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.826937 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.837066 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.850470 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.867944 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.880343 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.891816 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.907906 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.925142 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.932323 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.932371 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.932382 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.932403 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.932419 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:07Z","lastTransitionTime":"2026-02-20T00:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.946387 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.971425 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:07 crc kubenswrapper[4750]: I0220 00:09:07.993150 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:07Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.008057 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.023731 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.038032 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.039891 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.039966 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.039993 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.040026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.040049 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.158847 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.158912 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.158932 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.158957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.158975 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.261974 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.262023 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.262032 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.262046 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.262057 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.365532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.365607 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.365632 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.365668 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.365698 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.468493 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.468547 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.468558 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.468586 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.468602 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.490793 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 03:55:56.429876834 +0000 UTC Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.559278 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.559309 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:08 crc kubenswrapper[4750]: E0220 00:09:08.559488 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:08 crc kubenswrapper[4750]: E0220 00:09:08.559704 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.572834 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.572881 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.572908 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.572936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.572955 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.677525 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.677585 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.677603 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.677628 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.677645 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.781464 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.781538 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.781559 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.781583 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.781600 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.819871 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerStarted","Data":"1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.826220 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.826534 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.826571 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.842355 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.858535 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.858668 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.861406 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.872574 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.883758 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.883783 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.883795 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.883811 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.883824 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.889241 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.904091 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.918955 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.935697 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.956783 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.969935 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.984623 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.987636 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.987669 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.987677 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.987691 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.987700 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:08Z","lastTransitionTime":"2026-02-20T00:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:08 crc kubenswrapper[4750]: I0220 00:09:08.998526 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:08Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.012042 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.031229 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.042204 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.055852 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.067410 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.077141 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.086604 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.091490 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.091568 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.091588 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.091616 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.091641 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.103595 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.119645 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.140352 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.173516 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.195368 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.199395 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.199597 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.199623 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.199645 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.199679 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.214821 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.237838 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.259176 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.278524 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.295159 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.302585 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.302635 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.302651 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.302676 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.302721 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.405179 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.405258 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.405282 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.405314 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.405336 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.491012 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 16:33:14.22617182 +0000 UTC Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.509494 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.509560 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.509581 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.509608 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.509624 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.559281 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:09 crc kubenswrapper[4750]: E0220 00:09:09.559487 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.612817 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.612876 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.612894 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.612920 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.612938 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.716492 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.716568 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.716595 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.716629 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.716648 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.820659 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.820739 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.820763 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.820790 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.820809 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.834034 4750 generic.go:334] "Generic (PLEG): container finished" podID="02349b76-be13-4b84-a010-38792f5d9d86" containerID="1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0" exitCode=0 Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.834263 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.835580 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerDied","Data":"1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.856252 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.875360 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.889614 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.901921 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.912696 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.924287 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.924319 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.924338 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.924356 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.924370 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:09Z","lastTransitionTime":"2026-02-20T00:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.931159 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:09 crc kubenswrapper[4750]: I0220 00:09:09.994822 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:09Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.018569 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.027309 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.027342 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.027352 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.027366 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.027375 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.029099 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.043675 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.059283 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.076294 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.089763 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.103280 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.129769 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.129804 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.129814 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.129827 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.129837 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.231949 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.232000 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.232016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.232036 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.232051 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.336583 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.336614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.336622 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.336638 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.336647 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.439327 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.439359 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.439368 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.439384 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.439393 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.492093 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 03:07:11.215110877 +0000 UTC Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.541906 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.541933 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.541941 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.541954 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.541964 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.561455 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:10 crc kubenswrapper[4750]: E0220 00:09:10.561541 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.561596 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:10 crc kubenswrapper[4750]: E0220 00:09:10.561632 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.644075 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.644110 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.644143 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.644159 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.644172 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.748216 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.748296 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.748318 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.748347 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.748369 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.842583 4750 generic.go:334] "Generic (PLEG): container finished" podID="02349b76-be13-4b84-a010-38792f5d9d86" containerID="adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe" exitCode=0 Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.842671 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerDied","Data":"adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.842731 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.852393 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.852434 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.852445 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.852459 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.852472 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.857570 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.880582 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.920972 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.937092 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.951517 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.955884 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.955929 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.955944 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.955962 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.955974 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:10Z","lastTransitionTime":"2026-02-20T00:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.964370 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.978153 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:10 crc kubenswrapper[4750]: I0220 00:09:10.994387 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:10Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.006000 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.017913 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.029878 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.039442 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.041077 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.057901 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.058292 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.058369 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.058382 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.058406 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.058433 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.067833 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.077404 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.094640 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.104543 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.117028 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.130241 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.145655 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.157420 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.160867 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.160899 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.160910 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.160928 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.160939 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.204633 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.221730 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.238942 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.258953 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.264104 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.264174 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.264185 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.264208 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.264220 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.279262 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.296266 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.311956 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.367224 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.367277 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.367288 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.367312 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.367326 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.469903 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.469948 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.469957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.469972 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.469983 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.493166 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 05:51:18.213548058 +0000 UTC Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.559887 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:11 crc kubenswrapper[4750]: E0220 00:09:11.560004 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.572396 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.572449 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.572462 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.572482 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.572494 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.674690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.674727 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.674739 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.674756 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.674769 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.778593 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.778662 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.778682 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.778708 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.778725 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.849444 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/0.log" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.853610 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226" exitCode=1 Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.853707 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.854814 4750 scope.go:117] "RemoveContainer" containerID="e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.862426 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" event={"ID":"02349b76-be13-4b84-a010-38792f5d9d86","Type":"ContainerStarted","Data":"28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.879514 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.881702 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.881768 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.881787 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.881811 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.881829 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.900696 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.918586 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.938156 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.960985 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.982092 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:11Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.986187 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.986250 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.986288 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.986326 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:11 crc kubenswrapper[4750]: I0220 00:09:11.986350 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:11Z","lastTransitionTime":"2026-02-20T00:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.014310 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.046511 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"message\\\":\\\"0220 00:09:11.634498 6015 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:11.634617 6015 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:11.634637 6015 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:11.634821 6015 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:11.634851 6015 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:11.634856 6015 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:11.634884 6015 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 00:09:11.634892 6015 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:11.634910 6015 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:11.634925 6015 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 00:09:11.634930 6015 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:11.634933 6015 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 00:09:11.634942 6015 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:11.634953 6015 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:11.634975 6015 factory.go:656] Stopping watch factory\\\\nI0220 00:09:11.634994 6015 ovnkube.go:599] Stopped ovnkube\\\\nI0220 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.076991 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.089477 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.089536 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.089553 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.089577 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.089598 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.098498 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.122569 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.139904 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.159411 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.176071 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.192539 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.192616 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.192643 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.192672 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.192695 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.197954 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.211573 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.227040 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.236369 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.236438 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.236507 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.236545 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236656 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236703 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236709 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236726 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236737 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236757 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236797 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:28.23677261 +0000 UTC m=+52.431608669 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236828 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:28.236816051 +0000 UTC m=+52.431652280 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236832 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.236909 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:28.236890683 +0000 UTC m=+52.431726742 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.237228 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.237361 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:28.237347215 +0000 UTC m=+52.432183274 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.245231 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.258064 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.285247 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.294859 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.295046 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.295133 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.295241 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.295300 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.305851 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.320350 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.336577 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.337142 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.337332 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:09:28.337293419 +0000 UTC m=+52.532129488 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.359679 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"message\\\":\\\"0220 00:09:11.634498 6015 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:11.634617 6015 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:11.634637 6015 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:11.634821 6015 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:11.634851 6015 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:11.634856 6015 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:11.634884 6015 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 00:09:11.634892 6015 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:11.634910 6015 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:11.634925 6015 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 00:09:11.634930 6015 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:11.634933 6015 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 00:09:11.634942 6015 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:11.634953 6015 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:11.634975 6015 factory.go:656] Stopping watch factory\\\\nI0220 00:09:11.634994 6015 ovnkube.go:599] Stopped ovnkube\\\\nI0220 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.379331 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.393876 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.398001 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.398045 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.398057 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.398075 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.398091 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.412920 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.427335 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.493550 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 21:11:32.861222549 +0000 UTC Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.500859 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.500915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.500926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.500948 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.500960 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.559871 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.560036 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.560100 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.560384 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.603918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.603989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.604004 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.604026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.604038 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.706618 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.706679 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.706690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.706707 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.706719 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.809642 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.809709 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.809734 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.809760 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.809778 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.867735 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/1.log" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.868673 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/0.log" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.872589 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5" exitCode=1 Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.872662 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.872725 4750 scope.go:117] "RemoveContainer" containerID="e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.874256 4750 scope.go:117] "RemoveContainer" containerID="af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5" Feb 20 00:09:12 crc kubenswrapper[4750]: E0220 00:09:12.874578 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.891026 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.905736 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.915911 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.916047 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.916072 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.916102 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.916155 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:12Z","lastTransitionTime":"2026-02-20T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.923839 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.939175 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.953508 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.969387 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:12 crc kubenswrapper[4750]: I0220 00:09:12.982952 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:12.999976 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.019503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.019568 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.019585 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.019610 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.019634 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.026552 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.048383 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.072571 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"message\\\":\\\"0220 00:09:11.634498 6015 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:11.634617 6015 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:11.634637 6015 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:11.634821 6015 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:11.634851 6015 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:11.634856 6015 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:11.634884 6015 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 00:09:11.634892 6015 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:11.634910 6015 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:11.634925 6015 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 00:09:11.634930 6015 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:11.634933 6015 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 00:09:11.634942 6015 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:11.634953 6015 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:11.634975 6015 factory.go:656] Stopping watch factory\\\\nI0220 00:09:11.634994 6015 ovnkube.go:599] Stopped ovnkube\\\\nI0220 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.092092 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.105633 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.120456 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.123330 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.123405 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.123428 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.123458 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.123481 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.125929 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.125982 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.125996 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.126013 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.126026 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: E0220 00:09:13.143623 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.148343 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.148387 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.148401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.148419 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.148431 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: E0220 00:09:13.168916 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.173512 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.173561 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.173572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.173590 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.173604 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: E0220 00:09:13.191075 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.196878 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.196924 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.196935 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.196951 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.196962 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: E0220 00:09:13.210620 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.214780 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.214825 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.214840 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.214859 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.214871 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: E0220 00:09:13.228182 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:13Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:13 crc kubenswrapper[4750]: E0220 00:09:13.228363 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.232286 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.232323 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.232343 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.232365 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.232381 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.335589 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.335637 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.335647 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.335664 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.335676 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.437828 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.437886 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.437896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.437912 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.437922 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.493905 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 10:26:30.75563043 +0000 UTC Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.540852 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.540899 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.540911 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.540934 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.540946 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.559531 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:13 crc kubenswrapper[4750]: E0220 00:09:13.559696 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.644533 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.644600 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.644618 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.644641 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.644660 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.747818 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.747864 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.747872 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.747887 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.747896 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.855225 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.855307 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.855328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.855361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.855389 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.878637 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/1.log" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.958064 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.958390 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.958532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.958674 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:13 crc kubenswrapper[4750]: I0220 00:09:13.958795 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:13Z","lastTransitionTime":"2026-02-20T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.061626 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.061684 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.061696 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.061715 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.061733 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.164859 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.165108 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.165269 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.165406 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.165528 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.269443 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.269522 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.269533 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.269559 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.269571 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.371594 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.371949 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.372085 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.372308 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.372441 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.474916 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.474989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.475013 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.475040 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.475058 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.494542 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 21:19:50.590179912 +0000 UTC Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.559771 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.559774 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:14 crc kubenswrapper[4750]: E0220 00:09:14.560524 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:14 crc kubenswrapper[4750]: E0220 00:09:14.560724 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.577354 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.577436 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.577464 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.577494 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.577520 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.680938 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.681005 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.681028 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.681058 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.681081 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.785263 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.785344 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.785369 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.785400 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.785423 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.887834 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.887896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.887918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.887953 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.887989 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.990529 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.990594 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.990614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.990636 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:14 crc kubenswrapper[4750]: I0220 00:09:14.990654 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:14Z","lastTransitionTime":"2026-02-20T00:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.094019 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.094091 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.094111 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.094161 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.094178 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.197278 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.197337 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.197350 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.197370 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.197382 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.300106 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.300234 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.300258 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.300290 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.300353 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.403000 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.403031 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.403041 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.403056 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.403068 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.495711 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 21:33:35.630355501 +0000 UTC Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.506587 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.506648 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.506671 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.506716 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.506739 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.559259 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:15 crc kubenswrapper[4750]: E0220 00:09:15.559459 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.609899 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.609945 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.609965 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.609991 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.610009 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.713065 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.713162 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.713182 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.713206 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.713223 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.745479 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q"] Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.746370 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.749569 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.749968 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.767974 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.791592 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.805798 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.817586 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.817683 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.818288 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.818386 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.818504 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.822566 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.838774 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.852263 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.866105 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.879791 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6621b362-211d-4ebf-8ead-534d32b00b89-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.879928 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6621b362-211d-4ebf-8ead-534d32b00b89-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.880196 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6621b362-211d-4ebf-8ead-534d32b00b89-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.880448 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvkcj\" (UniqueName: \"kubernetes.io/projected/6621b362-211d-4ebf-8ead-534d32b00b89-kube-api-access-lvkcj\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.883818 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.902482 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.920321 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.921485 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.921554 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.921574 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.921640 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.921671 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:15Z","lastTransitionTime":"2026-02-20T00:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.935587 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.957650 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.977609 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.981157 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvkcj\" (UniqueName: \"kubernetes.io/projected/6621b362-211d-4ebf-8ead-534d32b00b89-kube-api-access-lvkcj\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.981480 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6621b362-211d-4ebf-8ead-534d32b00b89-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.981690 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6621b362-211d-4ebf-8ead-534d32b00b89-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.981955 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6621b362-211d-4ebf-8ead-534d32b00b89-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.982917 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6621b362-211d-4ebf-8ead-534d32b00b89-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.983515 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6621b362-211d-4ebf-8ead-534d32b00b89-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:15 crc kubenswrapper[4750]: I0220 00:09:15.992200 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6621b362-211d-4ebf-8ead-534d32b00b89-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.002144 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:15Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.008202 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvkcj\" (UniqueName: \"kubernetes.io/projected/6621b362-211d-4ebf-8ead-534d32b00b89-kube-api-access-lvkcj\") pod \"ovnkube-control-plane-749d76644c-2sb7q\" (UID: \"6621b362-211d-4ebf-8ead-534d32b00b89\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.024806 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.024852 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.024864 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.024883 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.024894 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.028303 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"message\\\":\\\"0220 00:09:11.634498 6015 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:11.634617 6015 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:11.634637 6015 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:11.634821 6015 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:11.634851 6015 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:11.634856 6015 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:11.634884 6015 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 00:09:11.634892 6015 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:11.634910 6015 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:11.634925 6015 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 00:09:11.634930 6015 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:11.634933 6015 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 00:09:11.634942 6015 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:11.634953 6015 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:11.634975 6015 factory.go:656] Stopping watch factory\\\\nI0220 00:09:11.634994 6015 ovnkube.go:599] Stopped ovnkube\\\\nI0220 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.075629 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" Feb 20 00:09:16 crc kubenswrapper[4750]: W0220 00:09:16.092142 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6621b362_211d_4ebf_8ead_534d32b00b89.slice/crio-9e8bc7d2506b651f1e9b4695d0aaafee8c09660198f4d64c278f454cf11ec83e WatchSource:0}: Error finding container 9e8bc7d2506b651f1e9b4695d0aaafee8c09660198f4d64c278f454cf11ec83e: Status 404 returned error can't find the container with id 9e8bc7d2506b651f1e9b4695d0aaafee8c09660198f4d64c278f454cf11ec83e Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.127678 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.127723 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.127738 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.127760 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.127774 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.230375 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.230423 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.230437 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.230459 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.230475 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.333429 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.333486 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.333505 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.333530 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.333547 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.436093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.436133 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.436141 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.436155 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.436164 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.496235 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 19:15:06.140980648 +0000 UTC Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.538497 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.538542 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.538556 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.538577 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.538595 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.559085 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.559255 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:16 crc kubenswrapper[4750]: E0220 00:09:16.559417 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:16 crc kubenswrapper[4750]: E0220 00:09:16.559541 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.576653 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.592109 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.606339 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.617993 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.633611 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.640004 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.640064 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.640079 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.640102 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.640142 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.647834 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.660765 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.677171 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.690354 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.703391 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.716247 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.735573 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.741970 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.742002 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.742015 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.742030 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.742040 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.756454 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.775229 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.799760 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"message\\\":\\\"0220 00:09:11.634498 6015 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:11.634617 6015 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:11.634637 6015 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:11.634821 6015 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:11.634851 6015 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:11.634856 6015 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:11.634884 6015 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 00:09:11.634892 6015 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:11.634910 6015 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:11.634925 6015 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 00:09:11.634930 6015 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:11.634933 6015 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 00:09:11.634942 6015 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:11.634953 6015 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:11.634975 6015 factory.go:656] Stopping watch factory\\\\nI0220 00:09:11.634994 6015 ovnkube.go:599] Stopped ovnkube\\\\nI0220 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.844161 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.844196 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.844209 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.844226 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.844237 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.895202 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" event={"ID":"6621b362-211d-4ebf-8ead-534d32b00b89","Type":"ContainerStarted","Data":"602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.895290 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" event={"ID":"6621b362-211d-4ebf-8ead-534d32b00b89","Type":"ContainerStarted","Data":"c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.895320 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" event={"ID":"6621b362-211d-4ebf-8ead-534d32b00b89","Type":"ContainerStarted","Data":"9e8bc7d2506b651f1e9b4695d0aaafee8c09660198f4d64c278f454cf11ec83e"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.916931 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.933700 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.947635 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.947971 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.948212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.948174 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.948419 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.948614 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:16Z","lastTransitionTime":"2026-02-20T00:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.964710 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.984145 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:16 crc kubenswrapper[4750]: I0220 00:09:16.999526 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:16Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.014537 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.025023 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.047185 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"message\\\":\\\"0220 00:09:11.634498 6015 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:11.634617 6015 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:11.634637 6015 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:11.634821 6015 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:11.634851 6015 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:11.634856 6015 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:11.634884 6015 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 00:09:11.634892 6015 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:11.634910 6015 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:11.634925 6015 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 00:09:11.634930 6015 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:11.634933 6015 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 00:09:11.634942 6015 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:11.634953 6015 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:11.634975 6015 factory.go:656] Stopping watch factory\\\\nI0220 00:09:11.634994 6015 ovnkube.go:599] Stopped ovnkube\\\\nI0220 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.051016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.051053 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.051073 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.051090 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.051099 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.066874 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.085967 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.102603 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.116985 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.130882 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.145657 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.153637 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.153678 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.153689 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.153708 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.153719 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.256556 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.256645 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.256664 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.256696 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.256713 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.272290 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-tcgh4"] Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.272778 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:17 crc kubenswrapper[4750]: E0220 00:09:17.272844 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.295928 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.316901 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.335567 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.350471 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.358753 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.358786 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.358796 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.358809 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.358818 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.368020 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.386492 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.395299 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.395404 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmtvx\" (UniqueName: \"kubernetes.io/projected/fac6ee29-99b8-42e7-844d-30e68cbecad3-kube-api-access-gmtvx\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.404218 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.433855 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e49d499da808ed46664c7883b1390af98850a409715250a4b73a5d8a4572b226\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"message\\\":\\\"0220 00:09:11.634498 6015 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:11.634617 6015 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:11.634637 6015 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:11.634821 6015 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:11.634851 6015 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:11.634856 6015 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:11.634884 6015 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 00:09:11.634892 6015 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:11.634910 6015 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:11.634925 6015 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 00:09:11.634930 6015 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:11.634933 6015 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 00:09:11.634942 6015 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:11.634953 6015 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:11.634975 6015 factory.go:656] Stopping watch factory\\\\nI0220 00:09:11.634994 6015 ovnkube.go:599] Stopped ovnkube\\\\nI0220 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.450793 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.460782 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.460837 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.460851 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.460870 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.460881 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.467509 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.482846 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.496390 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 23:30:43.335496089 +0000 UTC Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.496909 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.497069 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmtvx\" (UniqueName: \"kubernetes.io/projected/fac6ee29-99b8-42e7-844d-30e68cbecad3-kube-api-access-gmtvx\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:17 crc kubenswrapper[4750]: E0220 00:09:17.497200 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:17 crc kubenswrapper[4750]: E0220 00:09:17.497357 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:17.997325621 +0000 UTC m=+42.192161700 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.503948 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.519014 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmtvx\" (UniqueName: \"kubernetes.io/projected/fac6ee29-99b8-42e7-844d-30e68cbecad3-kube-api-access-gmtvx\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.521665 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.535538 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.549459 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.558435 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:17Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.559598 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:17 crc kubenswrapper[4750]: E0220 00:09:17.559746 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.563171 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.563278 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.563305 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.563337 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.563362 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.666254 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.666292 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.666301 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.666315 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.666328 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.769226 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.769283 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.769300 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.769323 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.769339 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.871931 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.871979 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.871990 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.872013 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.872027 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.974994 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.975051 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.975064 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.975083 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:17 crc kubenswrapper[4750]: I0220 00:09:17.975097 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:17Z","lastTransitionTime":"2026-02-20T00:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.003617 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:18 crc kubenswrapper[4750]: E0220 00:09:18.003758 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:18 crc kubenswrapper[4750]: E0220 00:09:18.003819 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:19.003801579 +0000 UTC m=+43.198637638 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.077940 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.077981 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.077993 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.078011 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.078023 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.181876 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.181944 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.181959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.181980 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.181995 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.285925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.285987 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.285998 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.286027 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.286050 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.388862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.388911 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.388923 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.388940 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.388952 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.492266 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.492328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.492345 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.492368 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.492387 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.496554 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:38:12.161607189 +0000 UTC Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.559670 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.559716 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.559857 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:18 crc kubenswrapper[4750]: E0220 00:09:18.560377 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:18 crc kubenswrapper[4750]: E0220 00:09:18.560358 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:18 crc kubenswrapper[4750]: E0220 00:09:18.560606 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.595853 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.596190 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.596270 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.596352 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.596422 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.700563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.701065 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.701084 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.701148 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.701175 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.803495 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.803547 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.803563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.803587 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.803606 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.905551 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.905599 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.905617 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.905649 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:18 crc kubenswrapper[4750]: I0220 00:09:18.905671 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:18Z","lastTransitionTime":"2026-02-20T00:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.008445 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.008503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.008519 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.008545 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.008562 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.016962 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:19 crc kubenswrapper[4750]: E0220 00:09:19.017101 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:19 crc kubenswrapper[4750]: E0220 00:09:19.017170 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:21.017156127 +0000 UTC m=+45.211992176 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.111024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.111088 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.111148 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.111181 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.111198 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.214424 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.214509 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.214530 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.214563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.214584 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.317478 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.317549 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.317572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.317602 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.317625 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.421103 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.421225 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.421251 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.421277 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.421299 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.497391 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 19:58:39.273410314 +0000 UTC Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.526840 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.526901 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.526919 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.526943 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.526961 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.559179 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:19 crc kubenswrapper[4750]: E0220 00:09:19.559328 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.629244 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.629316 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.629334 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.629357 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.629374 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.731737 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.731839 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.731863 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.731894 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.731920 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.835263 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.835339 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.835361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.835390 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.835408 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.938256 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.938305 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.938370 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.938393 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:19 crc kubenswrapper[4750]: I0220 00:09:19.938410 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:19Z","lastTransitionTime":"2026-02-20T00:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.041162 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.041223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.041245 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.041274 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.041300 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.144379 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.144425 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.144439 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.144465 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.144478 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.247728 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.247800 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.247821 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.247847 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.247864 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.351412 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.351475 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.351503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.351537 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.351561 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.454844 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.454920 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.454943 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.454970 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.454988 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.498320 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 02:42:09.472898774 +0000 UTC Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.558964 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.558983 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.559003 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.559105 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: E0220 00:09:20.559144 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.559160 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.559195 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.559219 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.559234 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: E0220 00:09:20.559470 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:20 crc kubenswrapper[4750]: E0220 00:09:20.559480 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.667587 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.667687 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.667743 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.667771 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.667863 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.771473 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.771552 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.771575 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.771606 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.771629 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.875333 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.875399 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.875417 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.875443 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.875459 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.978147 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.978224 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.978246 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.978277 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:20 crc kubenswrapper[4750]: I0220 00:09:20.978299 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:20Z","lastTransitionTime":"2026-02-20T00:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.037582 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:21 crc kubenswrapper[4750]: E0220 00:09:21.037794 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:21 crc kubenswrapper[4750]: E0220 00:09:21.037884 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:25.037858471 +0000 UTC m=+49.232694560 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.080889 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.081020 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.081040 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.081066 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.081083 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.184286 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.184343 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.184362 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.184385 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.184402 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.287214 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.287338 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.287379 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.287420 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.287444 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.390831 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.390877 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.390890 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.390908 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.390922 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.493925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.493989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.494006 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.494028 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.494046 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.498520 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 01:19:01.602395345 +0000 UTC Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.559601 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:21 crc kubenswrapper[4750]: E0220 00:09:21.559796 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.596912 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.597035 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.597074 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.597225 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.597253 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.700445 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.700508 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.700533 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.700563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.700584 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.803597 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.803645 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.803659 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.803684 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.803705 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.907300 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.907424 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.907442 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.907507 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:21 crc kubenswrapper[4750]: I0220 00:09:21.907526 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:21Z","lastTransitionTime":"2026-02-20T00:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.012053 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.012154 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.012180 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.012209 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.012225 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.115446 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.115513 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.115535 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.115569 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.115589 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.219107 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.219231 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.219248 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.219273 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.219292 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.323036 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.323098 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.323158 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.323191 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.323214 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.425981 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.426017 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.426027 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.426043 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.426054 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.499278 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 11:57:45.080245632 +0000 UTC Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.528813 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.528854 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.528864 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.528880 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.528893 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.559636 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:22 crc kubenswrapper[4750]: E0220 00:09:22.559812 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.560291 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:22 crc kubenswrapper[4750]: E0220 00:09:22.560400 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.560484 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:22 crc kubenswrapper[4750]: E0220 00:09:22.560569 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.632037 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.632104 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.632145 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.632170 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.632187 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.735060 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.735109 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.735154 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.735178 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.735197 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.838099 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.838203 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.838226 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.838255 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.838276 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.941361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.941446 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.941474 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.941506 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:22 crc kubenswrapper[4750]: I0220 00:09:22.941556 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:22Z","lastTransitionTime":"2026-02-20T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.045065 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.045111 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.045150 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.045170 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.045181 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.147808 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.147902 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.147928 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.147969 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.147996 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.251292 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.251349 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.251367 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.251396 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.251417 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.355045 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.355171 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.355201 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.355234 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.355257 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.459338 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.459423 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.459449 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.459482 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.459503 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.495936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.495991 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.496012 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.496053 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.496071 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.499973 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 23:10:02.906430735 +0000 UTC Feb 20 00:09:23 crc kubenswrapper[4750]: E0220 00:09:23.515626 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.520433 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.520479 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.520488 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.520508 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.520522 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: E0220 00:09:23.535720 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.540854 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.540904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.540919 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.540942 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.540959 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.558929 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:23 crc kubenswrapper[4750]: E0220 00:09:23.558929 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: E0220 00:09:23.559228 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.560320 4750 scope.go:117] "RemoveContainer" containerID="af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.566208 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.566248 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.566266 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.566294 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.566313 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.584764 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: E0220 00:09:23.587260 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.595860 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.595901 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.595915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.595938 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.595955 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.610050 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: E0220 00:09:23.620612 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: E0220 00:09:23.620770 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.623632 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.623689 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.623703 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.623728 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.623745 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.634070 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.669672 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.689271 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.716284 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.725965 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.726003 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.726016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.726034 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.726046 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.733343 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.752971 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.771080 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.787603 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.807171 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.823049 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.828312 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.828363 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.828380 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.828402 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.828417 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.844235 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.862650 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.878241 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.896924 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.922391 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/1.log" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.925141 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.925280 4750 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.931440 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.931499 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.931520 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.931549 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.931570 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:23Z","lastTransitionTime":"2026-02-20T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.945868 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.960424 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.976383 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:23 crc kubenswrapper[4750]: I0220 00:09:23.992335 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:23Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.011810 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.031756 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.033941 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.033987 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.033998 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.034016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.034027 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.052454 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.070343 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.093072 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.118683 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.137026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.137075 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.137089 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.137109 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.137135 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.140277 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.158376 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.180482 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.194832 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.209852 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.227641 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.240258 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.240317 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.240327 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.240347 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.240362 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.342559 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.342627 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.342639 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.342663 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.342681 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.446373 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.446433 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.446450 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.446472 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.446485 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.500870 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 05:58:01.36676701 +0000 UTC Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.549212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.549259 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.549267 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.549286 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.549300 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.559705 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.559738 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.559823 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:24 crc kubenswrapper[4750]: E0220 00:09:24.559924 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:24 crc kubenswrapper[4750]: E0220 00:09:24.560077 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:24 crc kubenswrapper[4750]: E0220 00:09:24.560210 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.652547 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.652614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.652637 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.652670 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.652690 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.756073 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.756170 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.756188 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.756213 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.756230 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.859207 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.859265 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.859283 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.859308 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.859329 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.930314 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/2.log" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.931192 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/1.log" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.934347 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f" exitCode=1 Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.934401 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.934447 4750 scope.go:117] "RemoveContainer" containerID="af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.935764 4750 scope.go:117] "RemoveContainer" containerID="43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f" Feb 20 00:09:24 crc kubenswrapper[4750]: E0220 00:09:24.936061 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.962614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.962662 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.962678 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.962704 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.962722 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:24Z","lastTransitionTime":"2026-02-20T00:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.963607 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:24 crc kubenswrapper[4750]: I0220 00:09:24.984395 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.003269 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.019954 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.044065 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.060882 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.066289 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.066337 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.066354 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.066380 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.066397 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.076368 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.088351 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.088816 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:25 crc kubenswrapper[4750]: E0220 00:09:25.089214 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:25 crc kubenswrapper[4750]: E0220 00:09:25.089313 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:33.089290497 +0000 UTC m=+57.284126556 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.104882 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.121220 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.138863 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.150661 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.169197 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.169269 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.169289 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.169320 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.169341 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.176478 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.200979 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.215739 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.230335 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:25Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.272057 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.272104 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.272137 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.272159 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.272175 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.374883 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.374965 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.374987 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.375013 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.375030 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.478679 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.478747 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.478765 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.478791 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.478807 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.501701 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 20:46:53.159016291 +0000 UTC Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.559471 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:25 crc kubenswrapper[4750]: E0220 00:09:25.559626 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.582026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.582092 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.582109 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.582162 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.582183 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.684438 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.684466 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.684474 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.684487 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.684496 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.787627 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.787664 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.787674 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.787692 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.787703 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.890746 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.890817 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.890839 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.890863 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.890880 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.941999 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/2.log" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.993868 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.993924 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.993941 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.993964 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:25 crc kubenswrapper[4750]: I0220 00:09:25.993982 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:25Z","lastTransitionTime":"2026-02-20T00:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.096458 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.096559 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.096582 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.096615 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.096648 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.199977 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.200039 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.200058 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.200083 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.200101 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.303315 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.303375 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.303394 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.303417 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.303434 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.406172 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.406221 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.406236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.406259 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.406276 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.502681 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 10:47:48.000921491 +0000 UTC Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.509389 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.509425 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.509434 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.509450 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.509464 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.559370 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.559473 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:26 crc kubenswrapper[4750]: E0220 00:09:26.559597 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:26 crc kubenswrapper[4750]: E0220 00:09:26.559755 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.559929 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:26 crc kubenswrapper[4750]: E0220 00:09:26.560166 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.582857 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.601380 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.611739 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.611779 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.611791 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.611809 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.611822 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.620842 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.656105 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.673014 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.686919 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.700825 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.714426 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.714498 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.714524 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.714555 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.714577 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.716683 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.732369 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.744029 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.756806 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.771359 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.786582 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.806176 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.817339 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.817387 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.817405 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.817429 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.817449 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.822103 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.837723 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.920901 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.922340 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.922401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.922423 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.922452 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.922474 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:26Z","lastTransitionTime":"2026-02-20T00:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.933885 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.944567 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.963483 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.979728 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:26 crc kubenswrapper[4750]: I0220 00:09:26.996515 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:26Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.011867 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.025704 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.025764 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.025783 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.025806 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.025823 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.029822 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.044014 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.057592 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.078902 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.096933 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.112874 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.127634 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.128772 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.128813 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.128834 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.128858 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.128874 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.149566 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.177253 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af59f60f232b0342f26468d53292d2acfc9888da5e3d3e079a4ba73c958528c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:12Z\\\",\\\"message\\\":\\\"b in node crc\\\\nI0220 00:09:12.826693 6180 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0220 00:09:12.826700 6180 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0220 00:09:12.826713 6180 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-q46qb\\\\nF0220 00:09:12.826720 6180 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:12Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:12.826722 6180 obj_retry.go:365] Adding new object: *\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.197546 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.216850 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:27Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.231212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.231345 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.231367 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.231391 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.231410 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.335157 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.335222 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.335239 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.335265 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.335283 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.438415 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.438479 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.438496 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.438522 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.438542 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.503305 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 12:39:22.947297858 +0000 UTC Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.541459 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.541520 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.541540 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.541566 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.541584 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.559156 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:27 crc kubenswrapper[4750]: E0220 00:09:27.559375 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.644957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.645031 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.645049 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.645078 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.645095 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.748820 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.748903 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.748932 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.748965 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.748987 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.852212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.852276 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.852293 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.852318 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.852338 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.955718 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.955781 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.955797 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.955822 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:27 crc kubenswrapper[4750]: I0220 00:09:27.955839 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:27Z","lastTransitionTime":"2026-02-20T00:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.059160 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.059310 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.059378 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.059403 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.059420 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.162509 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.162558 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.162570 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.162587 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.162600 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.265939 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.266001 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.266012 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.266033 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.266048 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.325900 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.325993 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.326045 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.326107 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326199 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326295 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:10:00.326271418 +0000 UTC m=+84.521107477 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326302 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326372 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326433 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326451 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326540 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326618 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326402 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:10:00.326372111 +0000 UTC m=+84.521208210 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326646 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326748 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:10:00.326689079 +0000 UTC m=+84.521525188 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.326884 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:10:00.326850883 +0000 UTC m=+84.521686932 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.369669 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.369740 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.369752 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.369792 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.369806 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.427668 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.427895 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:10:00.427852665 +0000 UTC m=+84.622688754 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.473758 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.473841 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.473860 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.473889 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.473907 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.504570 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 01:34:07.886117421 +0000 UTC Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.559677 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.559785 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.559899 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.559991 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.560245 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:28 crc kubenswrapper[4750]: E0220 00:09:28.560495 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.576917 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.577024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.577043 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.577110 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.577183 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.679627 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.679694 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.679713 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.679739 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.679757 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.783595 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.783872 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.783907 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.783942 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.783966 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.887967 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.888038 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.888053 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.888078 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.888095 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.991599 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.991689 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.991717 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.991750 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:28 crc kubenswrapper[4750]: I0220 00:09:28.991775 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:28Z","lastTransitionTime":"2026-02-20T00:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.094801 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.094870 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.094926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.094960 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.094981 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.197593 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.197654 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.197670 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.197694 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.197712 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.300406 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.300471 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.300488 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.300512 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.300535 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.403593 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.403636 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.403648 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.403666 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.403678 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.504710 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 09:45:56.951220705 +0000 UTC Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.505941 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.506020 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.506046 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.506074 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.506092 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.558834 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:29 crc kubenswrapper[4750]: E0220 00:09:29.559032 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.608630 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.608700 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.608725 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.608756 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.608779 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.711797 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.711864 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.711880 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.711904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.711921 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.815256 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.815309 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.815329 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.815364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.815382 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.917871 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.917918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.917928 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.917949 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:29 crc kubenswrapper[4750]: I0220 00:09:29.917961 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:29Z","lastTransitionTime":"2026-02-20T00:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.021802 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.021846 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.021855 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.021870 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.021879 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.124574 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.124608 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.124617 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.124633 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.124643 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.228510 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.228566 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.228583 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.228611 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.228628 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.320283 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.321537 4750 scope.go:117] "RemoveContainer" containerID="43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f" Feb 20 00:09:30 crc kubenswrapper[4750]: E0220 00:09:30.321784 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.332400 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.332495 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.332515 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.332543 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.332566 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.340301 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.357059 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.380651 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.394152 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.408840 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.421010 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.433930 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.436082 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.436131 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.436140 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.436156 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.436167 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.444718 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.460233 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.474859 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.489498 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.505421 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 14:30:40.966530099 +0000 UTC Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.506475 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.527263 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.539337 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.539389 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.539406 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.539431 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.539449 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.544472 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.559678 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.559743 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.559678 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:30 crc kubenswrapper[4750]: E0220 00:09:30.559838 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:30 crc kubenswrapper[4750]: E0220 00:09:30.559994 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:30 crc kubenswrapper[4750]: E0220 00:09:30.560219 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.564849 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.586791 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.600422 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:30Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.643061 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.643167 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.643194 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.643228 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.643250 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.746649 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.746718 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.746743 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.746772 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.746796 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.849399 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.849443 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.849454 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.849474 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.849486 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.953292 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.953363 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.953382 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.953407 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:30 crc kubenswrapper[4750]: I0220 00:09:30.953435 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:30Z","lastTransitionTime":"2026-02-20T00:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.056809 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.056876 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.056894 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.056921 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.056939 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.160170 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.160276 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.160300 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.160361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.160382 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.263535 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.263593 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.263606 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.263628 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.263643 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.366589 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.366653 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.366670 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.366695 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.366712 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.470644 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.470724 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.470739 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.470768 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.470785 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.505829 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 08:44:33.491179634 +0000 UTC Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.559629 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:31 crc kubenswrapper[4750]: E0220 00:09:31.559870 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.574424 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.574476 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.574494 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.574518 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.574537 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.677993 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.678063 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.678083 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.678149 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.678176 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.781771 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.782237 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.782390 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.782563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.782701 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.886068 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.886173 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.886193 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.886223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.886243 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.989616 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.989686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.989709 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.989744 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:31 crc kubenswrapper[4750]: I0220 00:09:31.989767 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:31Z","lastTransitionTime":"2026-02-20T00:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.092361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.092466 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.092485 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.092511 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.092529 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.196328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.196400 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.196426 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.196456 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.196480 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.299572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.299649 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.299669 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.299692 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.299711 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.403558 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.403660 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.403680 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.403734 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.403751 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.506261 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 12:14:03.252358221 +0000 UTC Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.507301 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.507401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.507422 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.507453 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.507479 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.559825 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.559899 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:32 crc kubenswrapper[4750]: E0220 00:09:32.560036 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.560144 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:32 crc kubenswrapper[4750]: E0220 00:09:32.560195 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:32 crc kubenswrapper[4750]: E0220 00:09:32.560509 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.611199 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.611252 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.611264 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.611282 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.611293 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.715620 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.715684 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.715707 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.715745 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.715769 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.819515 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.819569 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.819588 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.819614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.819632 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.923247 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.923309 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.923326 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.923351 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:32 crc kubenswrapper[4750]: I0220 00:09:32.923370 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:32Z","lastTransitionTime":"2026-02-20T00:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.026215 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.026284 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.026301 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.026331 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.026350 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.130332 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.130383 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.130401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.130428 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.130445 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.188487 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.188756 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.188922 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:09:49.188880092 +0000 UTC m=+73.383716301 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.233278 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.233336 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.233356 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.233384 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.233405 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.336412 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.336492 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.336520 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.336548 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.336567 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.440852 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.440927 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.440946 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.440977 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.440998 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.506758 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 23:17:53.976849804 +0000 UTC Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.544395 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.544445 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.544460 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.544484 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.544501 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.558983 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.559193 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.647555 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.647649 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.647671 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.647702 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.647722 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.740443 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.740513 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.740536 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.740562 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.740581 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.772613 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:33Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.778379 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.778440 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.778458 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.778485 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.778504 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.799409 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:33Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.804575 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.804645 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.804663 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.804689 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.804710 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.826636 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:33Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.832452 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.832516 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.832563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.832590 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.832609 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.849407 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:33Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.855517 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.855591 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.855603 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.855646 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.855660 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.873731 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:33Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:33 crc kubenswrapper[4750]: E0220 00:09:33.873890 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.876982 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.877043 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.877062 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.877091 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.877110 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.979858 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.979910 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.979929 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.979951 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:33 crc kubenswrapper[4750]: I0220 00:09:33.979962 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:33Z","lastTransitionTime":"2026-02-20T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.084096 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.084209 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.084227 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.084258 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.084276 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.187038 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.187110 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.187171 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.187204 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.187227 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.290503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.290561 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.290582 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.290614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.290637 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.393997 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.394055 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.394072 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.394098 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.394154 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.497574 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.497782 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.497823 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.497855 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.497877 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.507995 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 07:33:28.667181088 +0000 UTC Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.559997 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:34 crc kubenswrapper[4750]: E0220 00:09:34.560252 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.560801 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:34 crc kubenswrapper[4750]: E0220 00:09:34.560919 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.561053 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:34 crc kubenswrapper[4750]: E0220 00:09:34.561203 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.600686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.600753 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.600772 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.600797 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.600813 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.704310 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.704385 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.704408 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.704441 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.704476 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.807349 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.807414 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.807436 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.807462 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.807487 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.910194 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.910246 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.910258 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.910276 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:34 crc kubenswrapper[4750]: I0220 00:09:34.910290 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:34Z","lastTransitionTime":"2026-02-20T00:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.013586 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.013660 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.013679 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.013704 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.013721 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.116517 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.116595 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.116613 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.116637 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.116660 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.219687 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.219769 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.219786 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.219813 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.219831 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.322893 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.322976 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.323039 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.323073 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.323100 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.426831 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.426889 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.426908 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.426932 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.426951 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.509309 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 11:31:35.114876641 +0000 UTC Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.530686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.530992 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.531027 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.531069 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.531140 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.559055 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:35 crc kubenswrapper[4750]: E0220 00:09:35.559849 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.634650 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.635328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.635682 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.635939 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.636183 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.739464 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.739867 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.740065 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.740269 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.740415 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.843745 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.844087 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.844185 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.844262 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.844326 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.947313 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.947364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.947376 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.947398 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:35 crc kubenswrapper[4750]: I0220 00:09:35.947416 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:35Z","lastTransitionTime":"2026-02-20T00:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.049317 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.049368 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.049382 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.049402 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.049416 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.152819 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.152882 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.152903 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.152930 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.152949 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.256425 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.256518 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.256572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.256598 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.256648 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.359592 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.359644 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.359660 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.359687 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.359704 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.463136 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.463197 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.463215 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.463240 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.463256 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.510232 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 19:07:37.87541479 +0000 UTC Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.559793 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.559837 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.559793 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:36 crc kubenswrapper[4750]: E0220 00:09:36.560143 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:36 crc kubenswrapper[4750]: E0220 00:09:36.560406 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:36 crc kubenswrapper[4750]: E0220 00:09:36.561066 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.566360 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.566589 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.566699 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.566809 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.566911 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.610615 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.631577 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.654507 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.668578 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.669209 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.669348 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.669448 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.669557 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.669646 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.683519 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.695877 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.712057 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.725512 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.742836 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.759524 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.772783 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.772811 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.772819 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.772833 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.772843 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.778502 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.801178 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.829566 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.855717 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.876252 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.876343 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.876362 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.876395 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.876414 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.879219 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.905088 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.925265 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:36Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.978569 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.978612 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.978624 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.978642 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:36 crc kubenswrapper[4750]: I0220 00:09:36.978653 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:36Z","lastTransitionTime":"2026-02-20T00:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.082631 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.082683 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.082699 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.082723 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.082738 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.185756 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.185836 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.185861 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.185902 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.185925 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.288964 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.289385 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.289402 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.289425 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.289444 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.393078 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.393160 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.393176 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.393201 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.393219 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.495911 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.495957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.495972 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.495994 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.496010 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.510801 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 19:09:17.345404163 +0000 UTC Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.559976 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:37 crc kubenswrapper[4750]: E0220 00:09:37.560254 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.598977 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.599042 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.599063 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.599090 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.599110 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.702452 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.702505 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.702514 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.702535 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.702548 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.806957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.807037 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.807061 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.807093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.807149 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.910188 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.910259 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.910283 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.910307 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:37 crc kubenswrapper[4750]: I0220 00:09:37.910326 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:37Z","lastTransitionTime":"2026-02-20T00:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.012417 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.012467 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.012486 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.012728 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.012863 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.116019 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.116093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.116153 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.116186 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.116211 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.219617 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.219664 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.219680 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.219703 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.219719 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.322409 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.322475 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.322507 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.322533 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.322549 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.424358 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.424456 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.424484 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.424521 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.424545 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.511623 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 21:35:05.353052936 +0000 UTC Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.528032 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.528098 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.528137 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.528162 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.528182 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.559138 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.559227 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.559465 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:38 crc kubenswrapper[4750]: E0220 00:09:38.559685 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:38 crc kubenswrapper[4750]: E0220 00:09:38.559802 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:38 crc kubenswrapper[4750]: E0220 00:09:38.559927 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.630822 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.630857 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.630867 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.630895 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.630907 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.734821 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.734886 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.734906 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.734931 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.734949 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.837953 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.838013 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.838029 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.838054 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.838071 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.941819 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.941889 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.941905 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.941929 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:38 crc kubenswrapper[4750]: I0220 00:09:38.941946 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:38Z","lastTransitionTime":"2026-02-20T00:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.044975 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.045026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.045043 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.045066 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.045082 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.147727 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.147776 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.147792 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.147819 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.147836 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.251227 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.251277 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.251293 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.251317 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.251335 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.354386 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.354439 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.354455 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.354478 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.354496 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.457018 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.457063 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.457079 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.457101 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.457156 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.511831 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 07:58:52.55238465 +0000 UTC Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.559638 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:39 crc kubenswrapper[4750]: E0220 00:09:39.559810 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.560230 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.560264 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.560277 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.560294 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.560306 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.662809 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.662875 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.662900 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.662930 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.662954 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.765703 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.765750 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.765766 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.765789 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.765805 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.868483 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.868532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.868548 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.868570 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.868588 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.970807 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.970852 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.970870 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.970891 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:39 crc kubenswrapper[4750]: I0220 00:09:39.970906 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:39Z","lastTransitionTime":"2026-02-20T00:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.073572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.073619 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.073640 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.073671 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.073691 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.175670 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.175746 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.175767 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.175792 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.175811 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.279047 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.279091 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.279112 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.279192 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.279214 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.381730 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.381792 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.381808 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.381832 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.381849 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.484629 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.484678 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.484690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.484709 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.484722 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.512318 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 20:10:10.751717843 +0000 UTC Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.559794 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.559811 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.559832 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:40 crc kubenswrapper[4750]: E0220 00:09:40.559935 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:40 crc kubenswrapper[4750]: E0220 00:09:40.560070 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:40 crc kubenswrapper[4750]: E0220 00:09:40.560243 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.587077 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.587168 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.587187 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.587213 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.587233 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.690685 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.690766 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.690788 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.690820 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.690842 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.793978 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.794025 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.794035 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.794051 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.794063 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.897949 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.897996 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.898007 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.898026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:40 crc kubenswrapper[4750]: I0220 00:09:40.898038 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:40Z","lastTransitionTime":"2026-02-20T00:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.001264 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.001302 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.001333 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.001350 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.001360 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.103524 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.103570 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.103581 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.103599 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.103611 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.206205 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.206247 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.206256 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.206272 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.206282 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.308937 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.308977 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.308985 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.309002 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.309011 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.411958 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.412019 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.412038 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.412062 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.412080 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.512966 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 10:59:28.163499106 +0000 UTC Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.515560 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.515619 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.515642 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.515670 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.515688 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.559726 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:41 crc kubenswrapper[4750]: E0220 00:09:41.560491 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.560825 4750 scope.go:117] "RemoveContainer" containerID="43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f" Feb 20 00:09:41 crc kubenswrapper[4750]: E0220 00:09:41.561270 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.617868 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.617927 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.617944 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.617970 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.617990 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.721044 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.721097 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.721111 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.721148 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.721162 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.824312 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.824400 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.824424 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.824454 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.824477 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.930663 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.930816 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.930833 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.930850 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:41 crc kubenswrapper[4750]: I0220 00:09:41.930861 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:41Z","lastTransitionTime":"2026-02-20T00:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.033484 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.033536 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.033549 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.033570 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.033585 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.136419 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.136472 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.136485 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.136505 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.136517 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.239307 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.239346 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.239358 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.239374 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.239386 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.342862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.342908 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.342921 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.342939 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.342950 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.445506 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.445532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.445541 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.445554 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.445584 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.513165 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 21:56:29.087224773 +0000 UTC Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.549525 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.549586 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.549645 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.549679 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.549704 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.558830 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.558931 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:42 crc kubenswrapper[4750]: E0220 00:09:42.559001 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.559087 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:42 crc kubenswrapper[4750]: E0220 00:09:42.559226 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:42 crc kubenswrapper[4750]: E0220 00:09:42.559326 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.652920 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.652972 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.652983 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.653002 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.653021 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.755703 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.755749 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.755766 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.755786 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.755801 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.858515 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.858555 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.858564 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.858578 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.858587 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.962107 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.962185 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.962196 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.962214 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:42 crc kubenswrapper[4750]: I0220 00:09:42.962225 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:42Z","lastTransitionTime":"2026-02-20T00:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.065185 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.065251 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.065267 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.065290 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.065306 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.167900 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.167948 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.167974 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.167993 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.168005 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.270936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.270983 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.270991 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.271011 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.271021 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.374312 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.374388 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.374404 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.374428 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.374443 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.479528 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.479614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.479656 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.479701 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.479721 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.513338 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 13:28:23.267909902 +0000 UTC Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.559610 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:43 crc kubenswrapper[4750]: E0220 00:09:43.559742 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.582336 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.582359 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.582368 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.582381 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.582395 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.685102 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.685159 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.685171 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.685189 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.685200 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.788336 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.788395 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.788414 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.788436 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.788452 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.890784 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.890815 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.890825 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.890841 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.890851 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.993390 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.993455 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.993473 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.993499 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:43 crc kubenswrapper[4750]: I0220 00:09:43.993517 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:43Z","lastTransitionTime":"2026-02-20T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.096245 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.096313 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.096335 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.096362 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.096384 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.169355 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.169388 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.169397 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.169410 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.169421 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.187223 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:44Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.191823 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.191853 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.191862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.191874 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.191882 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.209242 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:44Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.213186 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.213216 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.213227 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.213241 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.213252 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.230544 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:44Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.235733 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.235765 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.235796 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.235813 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.235825 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.252968 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:44Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.256994 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.257032 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.257044 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.257063 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.257075 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.274389 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:44Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.274615 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.275792 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.275847 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.275864 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.275877 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.275887 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.377905 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.377950 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.377957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.377969 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.377977 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.480163 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.480194 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.480204 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.480219 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.480229 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.514301 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 12:51:17.118377525 +0000 UTC Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.559860 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.559930 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.559959 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.560048 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.560201 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:44 crc kubenswrapper[4750]: E0220 00:09:44.560359 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.582539 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.582572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.582585 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.582597 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.582608 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.685696 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.685731 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.685740 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.685755 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.685765 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.789508 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.789565 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.789582 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.789606 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.789622 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.892785 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.892832 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.892844 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.892863 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.892878 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.995572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.995615 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.995623 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.995638 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:44 crc kubenswrapper[4750]: I0220 00:09:44.995648 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:44Z","lastTransitionTime":"2026-02-20T00:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.099189 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.099241 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.099252 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.099269 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.099282 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.202540 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.202612 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.202629 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.202669 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.202682 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.304869 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.304955 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.304969 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.304989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.305002 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.408480 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.408544 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.408558 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.408582 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.408612 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.510846 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.510925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.510948 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.510977 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.510999 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.514987 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 02:07:37.122893609 +0000 UTC Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.559532 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:45 crc kubenswrapper[4750]: E0220 00:09:45.559703 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.613922 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.613969 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.613983 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.614000 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.614011 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.715731 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.715775 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.715789 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.715806 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.715817 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.818263 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.818372 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.818412 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.818435 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.818448 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.920785 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.920849 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.920868 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.920892 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:45 crc kubenswrapper[4750]: I0220 00:09:45.920915 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:45Z","lastTransitionTime":"2026-02-20T00:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.023800 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.023871 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.023884 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.023900 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.023934 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.126244 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.126294 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.126307 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.126325 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.126337 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.228503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.228566 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.228577 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.228613 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.228626 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.331354 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.331413 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.331429 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.331452 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.331469 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.433591 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.433630 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.433642 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.433658 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.433671 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.515788 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 18:01:34.357129623 +0000 UTC Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.536890 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.536953 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.536971 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.537000 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.537019 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.559880 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:46 crc kubenswrapper[4750]: E0220 00:09:46.560068 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.560280 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:46 crc kubenswrapper[4750]: E0220 00:09:46.560444 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.560536 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:46 crc kubenswrapper[4750]: E0220 00:09:46.560641 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.573912 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.587816 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.599956 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.610992 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.625430 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.635951 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.640462 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.640505 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.640519 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.640542 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.640556 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.653883 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.670463 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.687871 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.700490 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.714036 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.729234 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.743955 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.744014 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.744028 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.744064 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.744082 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.752346 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.771735 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.789775 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.803491 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.814865 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:46Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.846971 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.847024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.847037 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.847057 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.847076 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.949474 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.949545 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.949556 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.949578 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:46 crc kubenswrapper[4750]: I0220 00:09:46.949592 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:46Z","lastTransitionTime":"2026-02-20T00:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.051444 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.051516 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.051536 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.051568 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.051589 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.154825 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.154884 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.154896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.154915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.154927 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.257965 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.258035 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.258053 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.258084 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.258110 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.361502 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.361561 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.361573 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.361594 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.361608 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.464035 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.464152 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.464178 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.464207 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.464228 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.516598 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 11:18:39.115311131 +0000 UTC Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.559696 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:47 crc kubenswrapper[4750]: E0220 00:09:47.559895 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.566648 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.566703 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.566720 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.566741 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.566753 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.670477 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.670531 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.670544 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.670564 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.670576 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.773442 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.773493 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.773506 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.773522 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.773534 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.876034 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.876157 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.876176 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.876204 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.876256 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.978882 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.978919 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.978943 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.978957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:47 crc kubenswrapper[4750]: I0220 00:09:47.978965 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:47Z","lastTransitionTime":"2026-02-20T00:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.081038 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.081072 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.081080 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.081093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.081103 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.183266 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.183318 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.183329 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.183345 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.183357 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.285759 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.285819 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.285837 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.285861 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.285878 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.388859 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.388913 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.388925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.388947 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.388958 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.493641 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.493735 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.493745 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.493769 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.493782 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.516895 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 01:38:54.401152028 +0000 UTC Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.559559 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.559667 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:48 crc kubenswrapper[4750]: E0220 00:09:48.559777 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.559888 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:48 crc kubenswrapper[4750]: E0220 00:09:48.560641 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:48 crc kubenswrapper[4750]: E0220 00:09:48.560800 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.596429 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.596485 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.596497 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.596514 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.596525 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.700224 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.700279 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.700295 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.700313 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.700324 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.804227 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.804284 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.804298 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.804317 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.804329 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.907631 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.907760 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.907773 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.907792 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:48 crc kubenswrapper[4750]: I0220 00:09:48.907810 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:48Z","lastTransitionTime":"2026-02-20T00:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.010426 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.010487 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.010505 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.010531 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.010550 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.113051 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.113157 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.113180 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.113204 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.113222 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.215580 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.215621 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.215629 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.215645 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.215654 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.268643 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:49 crc kubenswrapper[4750]: E0220 00:09:49.268844 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:49 crc kubenswrapper[4750]: E0220 00:09:49.268908 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:10:21.2688904 +0000 UTC m=+105.463726459 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.318457 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.318655 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.318742 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.318827 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.318943 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.421534 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.421783 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.421894 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.421959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.422014 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.517387 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 03:10:14.800592037 +0000 UTC Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.524566 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.524632 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.524652 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.524678 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.524695 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.559070 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:49 crc kubenswrapper[4750]: E0220 00:09:49.559263 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.626729 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.627016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.627225 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.627261 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.627284 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.731218 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.731269 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.731292 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.731323 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.731347 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.835156 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.835210 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.835226 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.835250 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.835268 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.938298 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.938346 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.938361 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.938384 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:49 crc kubenswrapper[4750]: I0220 00:09:49.938403 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:49Z","lastTransitionTime":"2026-02-20T00:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.028808 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/0.log" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.028851 4750 generic.go:334] "Generic (PLEG): container finished" podID="1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030" containerID="bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6" exitCode=1 Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.028878 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerDied","Data":"bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.029454 4750 scope.go:117] "RemoveContainer" containerID="bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.043085 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.043141 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.043153 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.043168 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.043177 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.045977 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.068107 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:49Z\\\",\\\"message\\\":\\\"2026-02-20T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070\\\\n2026-02-20T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070 to /host/opt/cni/bin/\\\\n2026-02-20T00:09:04Z [verbose] multus-daemon started\\\\n2026-02-20T00:09:04Z [verbose] Readiness Indicator file check\\\\n2026-02-20T00:09:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.079147 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.091199 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.102310 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.120086 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.135994 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.146730 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.146793 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.146810 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.146835 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.146853 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.151424 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.165628 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.179609 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.191483 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.206294 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.235719 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.249961 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.250024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.250043 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.250069 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.250089 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.252534 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.268975 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.284438 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.300134 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:50Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.352707 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.352770 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.352783 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.352801 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.352816 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.454843 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.454900 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.454921 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.454995 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.455015 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.518064 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 19:03:42.352347166 +0000 UTC Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.558056 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.558105 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.558140 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.558161 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.558174 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.559040 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.559191 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:50 crc kubenswrapper[4750]: E0220 00:09:50.559297 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.559537 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:50 crc kubenswrapper[4750]: E0220 00:09:50.559630 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:50 crc kubenswrapper[4750]: E0220 00:09:50.559708 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.661602 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.661666 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.661686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.661715 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.661734 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.765284 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.765357 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.765374 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.765401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.765423 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.868233 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.868301 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.868320 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.868347 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.868366 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.970937 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.971000 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.971016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.971039 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:50 crc kubenswrapper[4750]: I0220 00:09:50.971054 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:50Z","lastTransitionTime":"2026-02-20T00:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.035329 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/0.log" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.035416 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerStarted","Data":"8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.053405 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.069517 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.073401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.073453 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.073475 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.073518 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.073537 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.083551 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.104146 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.123850 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:49Z\\\",\\\"message\\\":\\\"2026-02-20T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070\\\\n2026-02-20T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070 to /host/opt/cni/bin/\\\\n2026-02-20T00:09:04Z [verbose] multus-daemon started\\\\n2026-02-20T00:09:04Z [verbose] Readiness Indicator file check\\\\n2026-02-20T00:09:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.138493 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.160895 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.176751 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.176803 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.176824 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.176850 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.176866 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.181652 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.200188 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.216702 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.241678 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.274354 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.279482 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.279579 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.279604 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.279641 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.279668 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.291869 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.308168 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.328611 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.347937 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.364906 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:51Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.383069 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.383129 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.383140 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.383158 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.383171 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.487425 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.487523 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.487547 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.487579 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.487599 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.519138 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:01:18.794574966 +0000 UTC Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.559468 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:51 crc kubenswrapper[4750]: E0220 00:09:51.559676 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.591299 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.591356 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.591375 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.591401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.591419 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.695391 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.695479 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.695501 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.695526 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.695540 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.798079 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.798158 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.798177 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.798201 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.798219 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.900748 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.900820 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.900837 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.900863 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:51 crc kubenswrapper[4750]: I0220 00:09:51.900882 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:51Z","lastTransitionTime":"2026-02-20T00:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.004212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.004273 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.004289 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.004321 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.004339 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.106940 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.107026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.107061 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.107091 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.107153 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.210132 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.210205 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.210223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.210243 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.210254 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.312790 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.312850 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.312866 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.312889 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.312906 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.416365 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.416460 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.416479 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.416501 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.416519 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.519291 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.519365 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.519378 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.519287 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 04:44:02.430813625 +0000 UTC Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.519400 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.519494 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.559914 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.559993 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:52 crc kubenswrapper[4750]: E0220 00:09:52.560365 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.560758 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:52 crc kubenswrapper[4750]: E0220 00:09:52.560933 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:52 crc kubenswrapper[4750]: E0220 00:09:52.561743 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.562315 4750 scope.go:117] "RemoveContainer" containerID="43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.623384 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.623453 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.623475 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.623506 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.623528 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.727142 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.727201 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.727214 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.727237 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.727251 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.830529 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.830570 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.830587 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.830609 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.830625 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.933254 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.933314 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.933332 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.933358 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:52 crc kubenswrapper[4750]: I0220 00:09:52.933378 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:52Z","lastTransitionTime":"2026-02-20T00:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.036927 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.036997 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.037014 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.037041 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.037058 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.045350 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/2.log" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.049382 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.050146 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.068706 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.086889 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.112013 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.136554 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.140824 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.140899 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.140922 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.140952 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.140984 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.178105 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.207629 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:49Z\\\",\\\"message\\\":\\\"2026-02-20T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070\\\\n2026-02-20T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070 to /host/opt/cni/bin/\\\\n2026-02-20T00:09:04Z [verbose] multus-daemon started\\\\n2026-02-20T00:09:04Z [verbose] Readiness Indicator file check\\\\n2026-02-20T00:09:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.225589 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.241165 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.243328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.243371 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.243383 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.243402 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.243415 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.258071 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.274961 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.291265 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.305042 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.321620 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.341792 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.345737 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.345781 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.345793 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.345811 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.345825 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.355815 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.372170 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.409415 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:53Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.448439 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.448479 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.448491 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.448508 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.448519 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.520095 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 23:37:02.871842226 +0000 UTC Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.551587 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.551677 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.551701 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.551735 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.551758 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.559834 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:53 crc kubenswrapper[4750]: E0220 00:09:53.559989 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.654165 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.654222 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.654236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.654258 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.654274 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.758004 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.758062 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.758073 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.758097 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.758111 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.861788 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.861867 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.861893 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.861926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.861946 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.965800 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.965867 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.965882 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.965909 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:53 crc kubenswrapper[4750]: I0220 00:09:53.965925 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:53Z","lastTransitionTime":"2026-02-20T00:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.056965 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/3.log" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.058165 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/2.log" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.062842 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" exitCode=1 Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.062907 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.062999 4750 scope.go:117] "RemoveContainer" containerID="43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.064027 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.064350 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.078329 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.078387 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.078401 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.078424 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.078441 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.090359 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.112924 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.135213 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.167617 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43bec23789e9a2b0b67a489b0bb1cfe673b0dccfbdbbbb3d14c2f2a0ad9d798f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:24Z\\\",\\\"message\\\":\\\"=default\\\\nI0220 00:09:24.520229 6386 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-xjx4s\\\\nF0220 00:09:24.520396 6386 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:24Z is after 2025-08-24T17:21:41Z]\\\\nI0220 00:09:24.520214 6386 services_controller.go:443] Built service openshift-authentication/oauth-openshift LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.222\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.l\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:53Z\\\",\\\"message\\\":\\\"oval\\\\nI0220 00:09:53.587301 6777 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:53.587361 6777 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 00:09:53.587396 6777 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 00:09:53.587440 6777 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:53.588212 6777 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:53.588281 6777 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 00:09:53.588315 6777 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:53.588339 6777 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:53.588416 6777 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 00:09:53.588447 6777 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:53.588466 6777 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:53.588539 6777 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:53.588543 6777 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.182571 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.182633 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.182655 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.182685 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.182710 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.190411 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.211522 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.230170 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.250340 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.272410 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.285486 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.285531 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.285542 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.285557 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.285569 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.296818 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:49Z\\\",\\\"message\\\":\\\"2026-02-20T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070\\\\n2026-02-20T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070 to /host/opt/cni/bin/\\\\n2026-02-20T00:09:04Z [verbose] multus-daemon started\\\\n2026-02-20T00:09:04Z [verbose] Readiness Indicator file check\\\\n2026-02-20T00:09:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.313846 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.333274 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.350989 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.375579 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.388674 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.388759 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.388813 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.388838 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.388856 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.397415 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.421933 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.438074 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.480541 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.480606 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.480625 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.480685 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.480705 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.498937 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.503295 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.503354 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.503368 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.503393 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.503409 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.520859 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:13:06.354864425 +0000 UTC Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.523303 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.529020 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.529149 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.529178 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.529214 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.529239 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.546476 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.552758 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.552860 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.552880 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.552908 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.552928 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.559354 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.559537 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.559514 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.559722 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.559842 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.559962 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.570638 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.575324 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.575389 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.575409 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.575433 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.575454 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.592630 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:54Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:54 crc kubenswrapper[4750]: E0220 00:09:54.592917 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.595032 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.595072 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.595090 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.595112 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.595159 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.697637 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.697690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.697709 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.697734 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.697752 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.801058 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.801111 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.801142 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.801159 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.801173 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.903769 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.903837 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.903859 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.903888 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:54 crc kubenswrapper[4750]: I0220 00:09:54.903912 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:54Z","lastTransitionTime":"2026-02-20T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.007016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.007078 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.007096 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.007150 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.007169 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.069448 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/3.log" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.074615 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:09:55 crc kubenswrapper[4750]: E0220 00:09:55.074873 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.095904 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.109988 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.110342 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.110534 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.110732 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.110937 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.116794 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.140192 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.175351 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:53Z\\\",\\\"message\\\":\\\"oval\\\\nI0220 00:09:53.587301 6777 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:53.587361 6777 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 00:09:53.587396 6777 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 00:09:53.587440 6777 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:53.588212 6777 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:53.588281 6777 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 00:09:53.588315 6777 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:53.588339 6777 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:53.588416 6777 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 00:09:53.588447 6777 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:53.588466 6777 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:53.588539 6777 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:53.588543 6777 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.198599 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.213914 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.213972 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.213989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.214013 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.214031 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.221262 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.241464 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.261390 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.284395 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.307709 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:49Z\\\",\\\"message\\\":\\\"2026-02-20T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070\\\\n2026-02-20T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070 to /host/opt/cni/bin/\\\\n2026-02-20T00:09:04Z [verbose] multus-daemon started\\\\n2026-02-20T00:09:04Z [verbose] Readiness Indicator file check\\\\n2026-02-20T00:09:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.317252 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.317301 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.317317 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.317340 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.317357 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.327076 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.346573 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.365019 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.390215 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.411167 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.420793 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.420871 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.420922 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.420951 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.420969 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.428297 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.445708 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:55Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.522204 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 03:09:39.456687927 +0000 UTC Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.524805 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.524883 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.524905 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.524963 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.524982 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.559800 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:55 crc kubenswrapper[4750]: E0220 00:09:55.560021 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.628680 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.628746 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.628769 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.628795 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.628814 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.731816 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.731905 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.731930 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.731959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.731977 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.834918 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.834979 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.834997 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.835022 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.835039 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.938490 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.938595 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.938623 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.938657 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:55 crc kubenswrapper[4750]: I0220 00:09:55.938680 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:55Z","lastTransitionTime":"2026-02-20T00:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.042404 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.042608 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.042634 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.042665 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.042686 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.145661 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.145752 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.145777 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.145807 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.145830 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.249024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.249100 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.249181 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.249231 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.249259 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.352369 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.352431 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.352449 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.352474 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.352494 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.455484 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.455552 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.455572 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.455600 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.455623 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.522485 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 16:20:49.214482463 +0000 UTC Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.558838 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.558876 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.558898 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.558970 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:56 crc kubenswrapper[4750]: E0220 00:09:56.559080 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.558889 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.559235 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.559268 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.559289 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: E0220 00:09:56.559374 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:56 crc kubenswrapper[4750]: E0220 00:09:56.559261 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.579150 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"62d0755e-ca22-4187-aae7-65377cedcaa9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://705ba47dc57db056df715f852b5c8b55a0039d96855338540bc7c2667746d32f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tdtm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fk2zg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.602266 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d7605250a6224f9ec94054fd630483396096a4c3de109bab7288a8e6c5f7928\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.622071 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.642940 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dm8lb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:49Z\\\",\\\"message\\\":\\\"2026-02-20T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070\\\\n2026-02-20T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b12b3e4d-3d27-4d6f-8afe-082818e69070 to /host/opt/cni/bin/\\\\n2026-02-20T00:09:04Z [verbose] multus-daemon started\\\\n2026-02-20T00:09:04Z [verbose] Readiness Indicator file check\\\\n2026-02-20T00:09:49Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lzxkk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dm8lb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.660148 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-96tjn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18b1268a-b190-41c7-9746-a801bf44bd8b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11f521c9acd7cebde810886c612328d62ad3001e5cb4cb62daaefa88d6d1b555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlzf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-96tjn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.662796 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.662855 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.662880 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.662909 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.662933 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.679582 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6621b362-211d-4ebf-8ead-534d32b00b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36b873e9602e142fbf3a8d7354f1ce40d9222b25fb6cc717591d227ac707c33\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://602bee02eb487d3aaea6526195462a7cf796984d970e81805f317f9e3b00cf18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lvkcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2sb7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.695881 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac6ee29-99b8-42e7-844d-30e68cbecad3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtvx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:17Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tcgh4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.713624 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24a063e2-0310-4999-b09d-9837543cec0b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18e8fb662bdb2e090a84d3045a669c5466a7b1c71fb903072acfe6808c28f707\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://612e98225befb20f7defad03851d237a2654061beb40a362cad518b507f901d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://232d8d59b52256d7545f40fd5700ea75de392d7f6fbaf278208ee84af3042cf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://25dfd1318d766ba038ad1e451915755d156ddc52ed3d5932cb2e570fedd59ff4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.733598 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.753244 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.766234 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.766292 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.766314 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.766343 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.766366 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.770718 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.793608 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.813571 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.833387 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.863207 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02349b76-be13-4b84-a010-38792f5d9d86\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28677c4af5028c5cd28115d62644c41995bce3f86d183ca381dcb079f5a4182f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dca0b0f0d4e2d72c68902709946b5d28c1b14fc799b722196ee90205374e98e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8d492627285c67479765b69824279979911b26451cc12eb983823bdbfa62882\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a330ab7ab683efc79e991d263eea6cb6dc8d72fedf39ffcbb0c47c9d3fc88332\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f6f6f456a939dd420738589d540fcce0749a1846ad4f62bc4fb7f4a974cd8c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1eb6d9c31a178fcd8a476bccaa1bb9dbbf1cd8a3d578e863acfa55440acbd8c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adab0e604bbe82d042dde8cdad7bcb94a0df398605584d404a7a8d27f3911dbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cfpmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xjx4s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.869434 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.869519 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.869544 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.869577 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.869600 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.896245 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T00:09:53Z\\\",\\\"message\\\":\\\"oval\\\\nI0220 00:09:53.587301 6777 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 00:09:53.587361 6777 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0220 00:09:53.587396 6777 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0220 00:09:53.587440 6777 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0220 00:09:53.588212 6777 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0220 00:09:53.588225 6777 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 00:09:53.588281 6777 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 00:09:53.588315 6777 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 00:09:53.588339 6777 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0220 00:09:53.588416 6777 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0220 00:09:53.588447 6777 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 00:09:53.588466 6777 handler.go:208] Removed *v1.Node event handler 7\\\\nI0220 00:09:53.588539 6777 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0220 00:09:53.588543 6777 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hkqgt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.918202 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23d9ff9b-4699-4dd7-87ae-47e99dae502f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3727b12376751210d0b7004330254f5d6f1bc5f57ceb0d4162a0a68ba941c941\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d237d44cf88c364ea133c1b271e0c1f72d54db490286ac5eff720110ecab9ca0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecdbc0e6b9caa3fa7feefa52e75c5b47ea736fd3f28efdac74a791010856b73e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:09:56Z is after 2025-08-24T17:21:41Z" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.972260 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.972333 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.972356 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.972386 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:56 crc kubenswrapper[4750]: I0220 00:09:56.972407 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:56Z","lastTransitionTime":"2026-02-20T00:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.075268 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.075337 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.075359 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.075389 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.075413 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.178815 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.179327 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.179528 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.179685 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.179819 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.283589 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.283659 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.283682 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.283713 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.283736 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.387329 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.387382 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.387399 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.387426 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.387446 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.490916 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.490998 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.491024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.491054 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.491076 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.523253 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 18:56:51.14436779 +0000 UTC Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.559655 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:57 crc kubenswrapper[4750]: E0220 00:09:57.560090 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.594647 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.594702 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.594721 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.594750 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.594767 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.698005 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.698085 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.698107 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.698176 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.698201 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.801984 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.802065 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.802089 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.802158 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.802190 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.905488 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.905601 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.905630 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.905659 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:57 crc kubenswrapper[4750]: I0220 00:09:57.905681 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:57Z","lastTransitionTime":"2026-02-20T00:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.009434 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.009483 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.009502 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.009525 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.009542 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.112160 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.112256 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.112278 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.112310 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.112334 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.214739 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.214807 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.214824 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.214848 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.214866 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.318301 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.318392 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.318419 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.318456 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.318481 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.421467 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.421546 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.421569 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.421598 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.421619 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.523566 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 21:27:45.035663863 +0000 UTC Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.524272 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.524314 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.524331 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.524360 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.524381 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.559944 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.560015 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.560055 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:09:58 crc kubenswrapper[4750]: E0220 00:09:58.560143 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:09:58 crc kubenswrapper[4750]: E0220 00:09:58.560314 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:09:58 crc kubenswrapper[4750]: E0220 00:09:58.560470 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.627724 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.627788 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.627810 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.627840 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.627862 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.734744 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.734962 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.734990 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.735035 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.735059 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.838415 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.838463 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.838474 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.838493 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.838505 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.940784 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.940853 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.940871 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.940899 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:58 crc kubenswrapper[4750]: I0220 00:09:58.940917 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:58Z","lastTransitionTime":"2026-02-20T00:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.046057 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.046155 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.046175 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.046203 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.046230 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.149442 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.149497 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.149514 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.149537 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.149553 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.252391 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.252435 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.252444 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.252459 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.252468 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.355463 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.355508 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.355517 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.355532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.355542 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.459087 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.459186 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.459202 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.459227 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.459243 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.524027 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 00:48:17.025686478 +0000 UTC Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.559790 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:09:59 crc kubenswrapper[4750]: E0220 00:09:59.560000 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.561983 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.562032 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.562084 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.562108 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.562166 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.665475 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.665538 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.665556 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.665582 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.665599 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.769195 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.769261 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.769279 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.769304 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.769320 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.873084 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.873180 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.873204 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.873232 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.873250 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.976054 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.976105 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.976151 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.976176 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:09:59 crc kubenswrapper[4750]: I0220 00:09:59.976196 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:09:59Z","lastTransitionTime":"2026-02-20T00:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.079605 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.079671 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.079690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.079716 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.079735 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.183334 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.183404 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.183422 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.183449 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.183476 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.286613 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.286666 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.286682 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.286716 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.286733 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.390617 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.390671 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.390689 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.390714 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.390731 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.395605 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.395664 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.395713 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.395774 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.395861 4750 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.395888 4750 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.395957 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.396007 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.395958 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.396027 4750 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.396052 4750 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.396074 4750 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.395986 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:11:04.395952471 +0000 UTC m=+148.590788550 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.396176 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 00:11:04.396149767 +0000 UTC m=+148.590985846 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.396201 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 00:11:04.396187448 +0000 UTC m=+148.591023537 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.396224 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 00:11:04.396212268 +0000 UTC m=+148.591048347 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.493770 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.493835 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.493857 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.493889 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.493913 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.496552 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.496809 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:04.496785187 +0000 UTC m=+148.691621266 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.525045 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 05:25:53.651740141 +0000 UTC Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.559770 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.559874 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.559979 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.559885 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.560094 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:00 crc kubenswrapper[4750]: E0220 00:10:00.560226 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.596747 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.596800 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.596815 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.596837 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.596855 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.699910 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.699950 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.699959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.699973 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.699983 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.803006 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.803078 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.803100 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.803171 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.803195 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.904868 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.904913 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.904936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.904956 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:00 crc kubenswrapper[4750]: I0220 00:10:00.904971 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:00Z","lastTransitionTime":"2026-02-20T00:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.008745 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.008810 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.008827 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.008852 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.008871 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.111957 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.112006 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.112022 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.112044 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.112060 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.213999 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.214045 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.214057 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.214077 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.214093 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.324888 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.324932 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.324945 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.324965 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.324974 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.431677 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.432051 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.432494 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.432576 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.432596 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.525290 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 07:42:20.949267078 +0000 UTC Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.536314 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.536373 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.536453 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.536487 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.536553 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.558826 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:01 crc kubenswrapper[4750]: E0220 00:10:01.558992 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.639349 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.639896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.639984 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.640071 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.640187 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.743642 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.743697 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.743709 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.743728 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.743743 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.846075 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.846206 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.846227 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.846250 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.846268 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.949384 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.949726 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.949761 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.949890 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:01 crc kubenswrapper[4750]: I0220 00:10:01.949934 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:01Z","lastTransitionTime":"2026-02-20T00:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.053686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.053731 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.053747 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.053769 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.053783 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.155946 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.156028 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.156063 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.156093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.156164 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.259803 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.259876 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.259890 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.259936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.259947 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.362996 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.363111 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.363208 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.363415 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.363456 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.466254 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.466316 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.466332 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.466356 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.466375 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.526248 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 18:07:13.843754778 +0000 UTC Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.559342 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.559385 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.559395 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:02 crc kubenswrapper[4750]: E0220 00:10:02.559606 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:02 crc kubenswrapper[4750]: E0220 00:10:02.559813 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:02 crc kubenswrapper[4750]: E0220 00:10:02.559872 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.569153 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.569221 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.569244 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.569272 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.569294 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.671967 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.672037 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.672059 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.672089 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.672111 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.774654 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.774703 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.774720 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.774749 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.774774 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.877574 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.877664 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.877691 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.877727 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.877762 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.981407 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.981452 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.981466 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.981486 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:02 crc kubenswrapper[4750]: I0220 00:10:02.981502 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:02Z","lastTransitionTime":"2026-02-20T00:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.084342 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.084376 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.084387 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.084400 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.084409 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.186458 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.186501 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.186514 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.186530 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.186545 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.289069 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.289148 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.289162 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.289177 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.289186 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.392478 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.392521 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.392532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.392552 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.392564 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.495065 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.495090 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.495099 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.495148 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.495177 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.527663 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 06:37:56.907093214 +0000 UTC Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.559332 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:03 crc kubenswrapper[4750]: E0220 00:10:03.559500 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.598310 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.598393 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.598419 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.598449 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.598468 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.701481 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.701543 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.701559 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.701588 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.701605 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.805181 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.805241 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.805259 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.805282 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.805298 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.908470 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.908541 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.908563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.908591 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:03 crc kubenswrapper[4750]: I0220 00:10:03.908613 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:03Z","lastTransitionTime":"2026-02-20T00:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.012185 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.012331 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.012362 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.012392 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.012413 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.114913 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.114956 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.114972 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.114997 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.115013 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.217758 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.217812 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.217829 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.217850 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.217867 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.321101 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.321203 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.321225 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.321267 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.321288 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.425601 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.425763 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.425787 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.425810 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.425826 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.528003 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 02:09:18.352324037 +0000 UTC Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.528991 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.529042 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.529064 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.529095 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.529153 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.560015 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:04 crc kubenswrapper[4750]: E0220 00:10:04.560379 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.560441 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.560514 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:04 crc kubenswrapper[4750]: E0220 00:10:04.560635 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:04 crc kubenswrapper[4750]: E0220 00:10:04.560723 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.632341 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.632399 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.632415 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.632663 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:04 crc kubenswrapper[4750]: I0220 00:10:04.632703 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.736277 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.736341 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.736364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.736395 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.736423 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.839879 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.839920 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.839934 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.839951 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.839962 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.942971 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.943028 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.943040 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.943057 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.943068 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.963998 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.964029 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.964040 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.964054 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.964064 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: E0220 00:10:04.981218 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:04Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.986108 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.986182 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.986195 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.986223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:04.986238 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:04Z","lastTransitionTime":"2026-02-20T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: E0220 00:10:05.007098 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.011943 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.011976 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.011989 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.012009 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.012021 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: E0220 00:10:05.031194 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.035500 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.035528 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.035540 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.035558 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.035569 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: E0220 00:10:05.053437 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.057811 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.057857 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.057875 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.057897 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.057914 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: E0220 00:10:05.077750 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T00:10:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d9107560-2b75-4572-807e-08297e5eaea6\\\",\\\"systemUUID\\\":\\\"867946a5-592b-4cd2-93c8-f70a12380801\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:05Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:05 crc kubenswrapper[4750]: E0220 00:10:05.078217 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.079973 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.080016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.080033 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.080054 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.080070 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.182835 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.182909 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.182929 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.182952 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.182970 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.285778 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.285832 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.285849 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.285870 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.285886 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.388732 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.388787 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.388799 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.388817 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.388829 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.491849 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.491924 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.491947 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.491977 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.491998 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.529040 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 03:59:36.198678761 +0000 UTC Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.559698 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:05 crc kubenswrapper[4750]: E0220 00:10:05.559907 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.595328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.595392 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.595414 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.595443 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.595470 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.697760 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.697822 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.697840 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.697862 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.697879 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.801407 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.801464 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.801484 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.801509 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.801528 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.904482 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.904526 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.904536 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.904552 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:05 crc kubenswrapper[4750]: I0220 00:10:05.904565 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:05Z","lastTransitionTime":"2026-02-20T00:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.007356 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.007435 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.007459 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.007489 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.007511 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.109588 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.109661 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.109686 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.109716 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.109739 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.212962 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.213041 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.213064 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.213093 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.213111 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.315794 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.315875 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.315896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.315921 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.315938 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.418856 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.418912 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.418926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.418942 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.418955 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.521929 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.522010 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.522034 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.522067 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.522088 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.529631 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 13:13:03.219876208 +0000 UTC Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.559249 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.559255 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:06 crc kubenswrapper[4750]: E0220 00:10:06.559442 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.559671 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:06 crc kubenswrapper[4750]: E0220 00:10:06.560223 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:06 crc kubenswrapper[4750]: E0220 00:10:06.560434 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.560913 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:10:06 crc kubenswrapper[4750]: E0220 00:10:06.561198 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.580850 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f2fa9f1beb1e2cee0aeeb1abd5812eb12fdafc1052160fcb932dc2958a0f826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.599932 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-q46qb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46e36f94-f9ca-490e-9144-4d4fd2bf5eeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af91a3533ea6cf84ed36a1ca7b90cdac4b1242155053861f915399f67e537543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nd97j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:09:05Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-q46qb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.622536 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6b1e32-a6a0-4ed3-b542-9d898afde9dc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0220 00:08:50.418093 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 00:08:50.419138 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1092627244/tls.crt::/tmp/serving-cert-1092627244/tls.key\\\\\\\"\\\\nI0220 00:08:56.337270 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 00:08:56.339550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 00:08:56.339565 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 00:08:56.339597 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 00:08:56.339606 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 00:08:56.353907 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 00:08:56.353940 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353946 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 00:08:56.353951 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 00:08:56.353953 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 00:08:56.353956 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 00:08:56.353958 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 00:08:56.354012 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 00:08:56.357391 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:39Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T00:08:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T00:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T00:08:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.625376 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.625429 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.625447 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.625470 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.625487 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.646891 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://429e2f4e497b7317861ba0fa606284ba7875ef9c7a802cb9100132c577aeee46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6914e21ef8ecf698ac603c7ba758c5f8ee85bf631747514c92f395f6999980b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.671034 4750 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T00:08:56Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T00:10:06Z is after 2025-08-24T17:21:41Z" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.711352 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xjx4s" podStartSLOduration=65.711324762 podStartE2EDuration="1m5.711324762s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:06.710015106 +0000 UTC m=+90.904851195" watchObservedRunningTime="2026-02-20 00:10:06.711324762 +0000 UTC m=+90.906160851" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.728269 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.728347 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.728371 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.728403 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.728426 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.789085 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.789062745 podStartE2EDuration="1m9.789062745s" podCreationTimestamp="2026-02-20 00:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:06.765010897 +0000 UTC m=+90.959846956" watchObservedRunningTime="2026-02-20 00:10:06.789062745 +0000 UTC m=+90.983898804" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.789269 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podStartSLOduration=65.789262971 podStartE2EDuration="1m5.789262971s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:06.789046985 +0000 UTC m=+90.983883044" watchObservedRunningTime="2026-02-20 00:10:06.789262971 +0000 UTC m=+90.984099040" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.831388 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.831447 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.831466 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.831491 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.831510 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.859837 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-dm8lb" podStartSLOduration=65.859814103 podStartE2EDuration="1m5.859814103s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:06.848592295 +0000 UTC m=+91.043428354" watchObservedRunningTime="2026-02-20 00:10:06.859814103 +0000 UTC m=+91.054650162" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.873801 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-96tjn" podStartSLOduration=65.873779224 podStartE2EDuration="1m5.873779224s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:06.861334753 +0000 UTC m=+91.056170812" watchObservedRunningTime="2026-02-20 00:10:06.873779224 +0000 UTC m=+91.068615283" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.874817 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2sb7q" podStartSLOduration=64.874809321 podStartE2EDuration="1m4.874809321s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:06.872256093 +0000 UTC m=+91.067092182" watchObservedRunningTime="2026-02-20 00:10:06.874809321 +0000 UTC m=+91.069645380" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.903546 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.903527333 podStartE2EDuration="40.903527333s" podCreationTimestamp="2026-02-20 00:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:06.90304458 +0000 UTC m=+91.097880639" watchObservedRunningTime="2026-02-20 00:10:06.903527333 +0000 UTC m=+91.098363392" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.934250 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.934500 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.934588 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.934847 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:06 crc kubenswrapper[4750]: I0220 00:10:06.934914 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:06Z","lastTransitionTime":"2026-02-20T00:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.036959 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.037830 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.037939 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.038011 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.038076 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.140607 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.141038 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.141236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.141420 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.141569 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.244486 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.244556 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.244575 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.244608 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.244655 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.348546 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.348674 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.348697 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.348722 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.348739 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.451615 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.451685 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.451702 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.451730 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.451749 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.530782 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 15:14:41.864699271 +0000 UTC Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.554290 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.554346 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.554364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.554389 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.554407 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.559530 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:07 crc kubenswrapper[4750]: E0220 00:10:07.559667 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.657938 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.658025 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.658047 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.658073 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.658090 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.761083 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.761182 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.761212 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.761236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.761255 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.864785 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.864834 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.864851 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.864874 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.864891 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.968328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.968391 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.968413 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.968441 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:07 crc kubenswrapper[4750]: I0220 00:10:07.968463 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:07Z","lastTransitionTime":"2026-02-20T00:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.071266 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.071315 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.071331 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.071351 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.071366 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.174160 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.174223 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.174243 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.174267 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.174284 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.276696 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.276775 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.276798 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.276826 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.276843 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.380916 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.380986 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.381008 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.381035 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.381057 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.484423 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.484498 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.484520 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.484549 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.484572 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.531562 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 11:29:24.615128617 +0000 UTC Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.559367 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.559423 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.559640 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:08 crc kubenswrapper[4750]: E0220 00:10:08.559778 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:08 crc kubenswrapper[4750]: E0220 00:10:08.559907 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:08 crc kubenswrapper[4750]: E0220 00:10:08.560108 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.576081 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.587286 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.587352 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.587364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.587384 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.587397 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.689627 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.689693 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.689715 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.689749 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.689773 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.792593 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.792647 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.792663 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.792685 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.792706 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.895927 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.895985 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.896001 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.896026 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.896043 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.998915 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.999024 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.999043 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.999069 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:08 crc kubenswrapper[4750]: I0220 00:10:08.999088 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:08Z","lastTransitionTime":"2026-02-20T00:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.101934 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.101995 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.102017 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.102042 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.102060 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.204592 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.204665 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.204690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.204724 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.204748 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.307308 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.307348 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.307359 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.307374 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.307386 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.409848 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.409896 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.410000 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.410027 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.410039 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.512554 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.512585 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.512596 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.512609 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.512618 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.532583 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 23:39:47.866345316 +0000 UTC Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.559109 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:09 crc kubenswrapper[4750]: E0220 00:10:09.559357 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.614926 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.614980 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.615001 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.615020 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.615034 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.718293 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.718354 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.718371 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.718395 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.718411 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.821762 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.821849 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.821875 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.821908 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.821931 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.925444 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.925506 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.925538 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.925568 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:09 crc kubenswrapper[4750]: I0220 00:10:09.925588 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:09Z","lastTransitionTime":"2026-02-20T00:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.027876 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.027952 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.027982 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.028013 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.028034 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.135774 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.135834 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.135853 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.135877 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.135894 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.241225 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.241295 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.241318 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.241349 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.241368 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.345405 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.345503 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.345573 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.345597 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.345661 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.447886 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.447948 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.447964 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.447988 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.448007 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.533168 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 08:04:54.88156788 +0000 UTC Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.550799 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.550841 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.550850 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.550865 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.550875 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.559426 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.559438 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.559547 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:10 crc kubenswrapper[4750]: E0220 00:10:10.559664 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:10 crc kubenswrapper[4750]: E0220 00:10:10.559889 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:10 crc kubenswrapper[4750]: E0220 00:10:10.560068 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.654069 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.654182 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.654208 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.654249 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.654267 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.756812 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.756870 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.756893 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.756921 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.756943 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.859655 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.859717 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.859751 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.859785 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.859809 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.962742 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.962815 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.962837 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.962884 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:10 crc kubenswrapper[4750]: I0220 00:10:10.962904 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:10Z","lastTransitionTime":"2026-02-20T00:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.064861 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.064906 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.064920 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.064936 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.064946 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.168416 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.168741 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.168755 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.168771 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.168785 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.271405 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.271456 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.271464 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.271476 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.271485 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.384282 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.384325 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.384342 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.384364 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.384380 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.486970 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.487027 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.487043 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.487066 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.487082 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.533769 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 19:16:18.580398635 +0000 UTC Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.559330 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:11 crc kubenswrapper[4750]: E0220 00:10:11.559739 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.590293 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.590357 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.590374 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.590399 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.590440 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.692947 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.692990 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.693003 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.693019 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.693030 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.795555 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.795582 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.795591 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.795604 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.795612 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.898512 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.898586 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.898603 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.898632 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:11 crc kubenswrapper[4750]: I0220 00:10:11.898649 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:11Z","lastTransitionTime":"2026-02-20T00:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.001910 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.001968 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.001984 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.002007 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.002024 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.105156 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.105218 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.105236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.105261 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.105280 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.207833 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.207904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.207922 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.207947 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.207973 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.310970 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.311094 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.311140 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.311163 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.311181 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.413733 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.413805 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.413854 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.413878 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.413895 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.528736 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.528801 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.528818 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.528842 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.528859 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.534662 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 11:55:02.522931448 +0000 UTC Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.559255 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:12 crc kubenswrapper[4750]: E0220 00:10:12.559396 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.559571 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:12 crc kubenswrapper[4750]: E0220 00:10:12.559630 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.559782 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:12 crc kubenswrapper[4750]: E0220 00:10:12.559865 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.631367 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.631410 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.631426 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.631447 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.631462 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.733632 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.733667 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.733677 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.733692 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.733702 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.836921 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.836976 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.836992 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.837016 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.837033 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.939432 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.939476 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.939488 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.939506 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:12 crc kubenswrapper[4750]: I0220 00:10:12.939519 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:12Z","lastTransitionTime":"2026-02-20T00:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.043177 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.043236 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.043252 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.043273 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.043289 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.146331 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.146415 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.146449 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.146466 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.146478 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.249773 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.249843 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.249868 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.249902 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.249928 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.354074 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.354213 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.354246 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.354277 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.354301 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.456739 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.456803 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.456820 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.456844 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.456866 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.535712 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 10:35:07.305666433 +0000 UTC Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.559735 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:13 crc kubenswrapper[4750]: E0220 00:10:13.559921 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.560532 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.560595 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.560614 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.560639 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.560657 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.664385 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.664443 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.664461 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.664484 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.664500 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.768835 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.768898 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.768914 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.769023 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.769043 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.872778 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.872875 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.872925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.872950 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.872968 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.976927 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.976981 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.976999 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.977022 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:13 crc kubenswrapper[4750]: I0220 00:10:13.977041 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:13Z","lastTransitionTime":"2026-02-20T00:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.080632 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.080726 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.080781 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.080808 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.080827 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.184441 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.184505 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.184522 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.184545 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.184564 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.287690 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.287798 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.287817 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.287843 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.287861 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.391425 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.391508 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.391534 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.391565 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.391587 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.494491 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.494547 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.494560 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.494577 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.494590 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.536433 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 05:36:28.844220997 +0000 UTC Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.559581 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.559775 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:14 crc kubenswrapper[4750]: E0220 00:10:14.559886 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.559945 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:14 crc kubenswrapper[4750]: E0220 00:10:14.560049 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:14 crc kubenswrapper[4750]: E0220 00:10:14.560222 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.597592 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.597656 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.597669 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.597691 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.597708 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.700836 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.700907 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.700925 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.700956 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.700975 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.804422 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.804493 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.804512 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.804539 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.804559 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.907655 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.907716 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.907726 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.907746 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:14 crc kubenswrapper[4750]: I0220 00:10:14.907759 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:14Z","lastTransitionTime":"2026-02-20T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.011237 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.011296 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.011309 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.011330 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.011348 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:15Z","lastTransitionTime":"2026-02-20T00:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.114578 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.114637 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.114647 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.114704 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.114717 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:15Z","lastTransitionTime":"2026-02-20T00:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.217108 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.217174 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.217185 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.217204 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.217216 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:15Z","lastTransitionTime":"2026-02-20T00:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.320506 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.320539 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.320548 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.320563 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.320573 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:15Z","lastTransitionTime":"2026-02-20T00:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.422904 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.422946 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.422962 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.422983 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.423001 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:15Z","lastTransitionTime":"2026-02-20T00:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.446274 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.446328 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.446346 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.446366 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.446382 4750 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T00:10:15Z","lastTransitionTime":"2026-02-20T00:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.510536 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr"] Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.511113 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.514979 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.515012 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.516399 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.518461 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.536978 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 15:51:22.219016893 +0000 UTC Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.537151 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.548829 4750 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.559423 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:15 crc kubenswrapper[4750]: E0220 00:10:15.559596 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.560298 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1704cca3-ee56-43df-9815-473f52fc317d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.560382 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1704cca3-ee56-43df-9815-473f52fc317d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.560421 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1704cca3-ee56-43df-9815-473f52fc317d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.560469 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1704cca3-ee56-43df-9815-473f52fc317d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.560514 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1704cca3-ee56-43df-9815-473f52fc317d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.575026 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-q46qb" podStartSLOduration=74.574964931 podStartE2EDuration="1m14.574964931s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:15.571333785 +0000 UTC m=+99.766169894" watchObservedRunningTime="2026-02-20 00:10:15.574964931 +0000 UTC m=+99.769801010" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.598522 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.598488565 podStartE2EDuration="1m19.598488565s" podCreationTimestamp="2026-02-20 00:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:15.597894409 +0000 UTC m=+99.792730508" watchObservedRunningTime="2026-02-20 00:10:15.598488565 +0000 UTC m=+99.793324654" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.624603 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=7.624569638 podStartE2EDuration="7.624569638s" podCreationTimestamp="2026-02-20 00:10:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:15.622870462 +0000 UTC m=+99.817706541" watchObservedRunningTime="2026-02-20 00:10:15.624569638 +0000 UTC m=+99.819405727" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.661595 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1704cca3-ee56-43df-9815-473f52fc317d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.661645 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1704cca3-ee56-43df-9815-473f52fc317d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.661691 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1704cca3-ee56-43df-9815-473f52fc317d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.661720 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1704cca3-ee56-43df-9815-473f52fc317d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.661731 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1704cca3-ee56-43df-9815-473f52fc317d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.661812 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1704cca3-ee56-43df-9815-473f52fc317d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.661882 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1704cca3-ee56-43df-9815-473f52fc317d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.663272 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1704cca3-ee56-43df-9815-473f52fc317d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.673200 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1704cca3-ee56-43df-9815-473f52fc317d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.680017 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1704cca3-ee56-43df-9815-473f52fc317d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2kfwr\" (UID: \"1704cca3-ee56-43df-9815-473f52fc317d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:15 crc kubenswrapper[4750]: I0220 00:10:15.840536 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" Feb 20 00:10:16 crc kubenswrapper[4750]: I0220 00:10:16.183376 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" event={"ID":"1704cca3-ee56-43df-9815-473f52fc317d","Type":"ContainerStarted","Data":"b4c230f9416fa82a6389abdc6d898a0b909f3ced57cc78f0788736ff9009c7d6"} Feb 20 00:10:16 crc kubenswrapper[4750]: I0220 00:10:16.183768 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" event={"ID":"1704cca3-ee56-43df-9815-473f52fc317d","Type":"ContainerStarted","Data":"9674b3d8b8dcb0b31ed59f521750724f907d22ed54227139232338a0c5e4bb60"} Feb 20 00:10:16 crc kubenswrapper[4750]: I0220 00:10:16.205655 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2kfwr" podStartSLOduration=75.205619771 podStartE2EDuration="1m15.205619771s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:16.204044438 +0000 UTC m=+100.398880547" watchObservedRunningTime="2026-02-20 00:10:16.205619771 +0000 UTC m=+100.400455860" Feb 20 00:10:16 crc kubenswrapper[4750]: I0220 00:10:16.558952 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:16 crc kubenswrapper[4750]: I0220 00:10:16.558977 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:16 crc kubenswrapper[4750]: E0220 00:10:16.560968 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:16 crc kubenswrapper[4750]: I0220 00:10:16.561091 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:16 crc kubenswrapper[4750]: E0220 00:10:16.561317 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:16 crc kubenswrapper[4750]: E0220 00:10:16.561793 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:16 crc kubenswrapper[4750]: I0220 00:10:16.583204 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 20 00:10:17 crc kubenswrapper[4750]: I0220 00:10:17.559240 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:17 crc kubenswrapper[4750]: E0220 00:10:17.559791 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:17 crc kubenswrapper[4750]: I0220 00:10:17.560299 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:10:17 crc kubenswrapper[4750]: E0220 00:10:17.560549 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:10:18 crc kubenswrapper[4750]: I0220 00:10:18.559340 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:18 crc kubenswrapper[4750]: I0220 00:10:18.559369 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:18 crc kubenswrapper[4750]: I0220 00:10:18.559369 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:18 crc kubenswrapper[4750]: E0220 00:10:18.559513 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:18 crc kubenswrapper[4750]: E0220 00:10:18.559703 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:18 crc kubenswrapper[4750]: E0220 00:10:18.559798 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:19 crc kubenswrapper[4750]: I0220 00:10:19.559239 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:19 crc kubenswrapper[4750]: E0220 00:10:19.559462 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:20 crc kubenswrapper[4750]: I0220 00:10:20.559969 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:20 crc kubenswrapper[4750]: I0220 00:10:20.560077 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:20 crc kubenswrapper[4750]: E0220 00:10:20.561427 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:20 crc kubenswrapper[4750]: I0220 00:10:20.560145 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:20 crc kubenswrapper[4750]: E0220 00:10:20.561563 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:20 crc kubenswrapper[4750]: E0220 00:10:20.561633 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:21 crc kubenswrapper[4750]: I0220 00:10:21.332325 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:21 crc kubenswrapper[4750]: E0220 00:10:21.332538 4750 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:10:21 crc kubenswrapper[4750]: E0220 00:10:21.332604 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs podName:fac6ee29-99b8-42e7-844d-30e68cbecad3 nodeName:}" failed. No retries permitted until 2026-02-20 00:11:25.332582346 +0000 UTC m=+169.527418425 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs") pod "network-metrics-daemon-tcgh4" (UID: "fac6ee29-99b8-42e7-844d-30e68cbecad3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 00:10:21 crc kubenswrapper[4750]: I0220 00:10:21.559291 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:21 crc kubenswrapper[4750]: E0220 00:10:21.560075 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:22 crc kubenswrapper[4750]: I0220 00:10:22.559328 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:22 crc kubenswrapper[4750]: I0220 00:10:22.559451 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:22 crc kubenswrapper[4750]: E0220 00:10:22.559553 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:22 crc kubenswrapper[4750]: I0220 00:10:22.559652 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:22 crc kubenswrapper[4750]: E0220 00:10:22.559763 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:22 crc kubenswrapper[4750]: E0220 00:10:22.559957 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:23 crc kubenswrapper[4750]: I0220 00:10:23.559727 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:23 crc kubenswrapper[4750]: E0220 00:10:23.559969 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:24 crc kubenswrapper[4750]: I0220 00:10:24.559554 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:24 crc kubenswrapper[4750]: I0220 00:10:24.559745 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:24 crc kubenswrapper[4750]: E0220 00:10:24.559883 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:24 crc kubenswrapper[4750]: E0220 00:10:24.560172 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:24 crc kubenswrapper[4750]: I0220 00:10:24.560287 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:24 crc kubenswrapper[4750]: E0220 00:10:24.560438 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:25 crc kubenswrapper[4750]: I0220 00:10:25.559062 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:25 crc kubenswrapper[4750]: E0220 00:10:25.559298 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:26 crc kubenswrapper[4750]: I0220 00:10:26.559385 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:26 crc kubenswrapper[4750]: I0220 00:10:26.559496 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:26 crc kubenswrapper[4750]: E0220 00:10:26.559624 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:26 crc kubenswrapper[4750]: I0220 00:10:26.559694 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:26 crc kubenswrapper[4750]: E0220 00:10:26.561756 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:26 crc kubenswrapper[4750]: E0220 00:10:26.561861 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:26 crc kubenswrapper[4750]: I0220 00:10:26.614304 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=10.61426527 podStartE2EDuration="10.61426527s" podCreationTimestamp="2026-02-20 00:10:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:26.606313629 +0000 UTC m=+110.801149758" watchObservedRunningTime="2026-02-20 00:10:26.61426527 +0000 UTC m=+110.809101359" Feb 20 00:10:27 crc kubenswrapper[4750]: I0220 00:10:27.559799 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:27 crc kubenswrapper[4750]: E0220 00:10:27.560051 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:28 crc kubenswrapper[4750]: I0220 00:10:28.559473 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:28 crc kubenswrapper[4750]: I0220 00:10:28.559553 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:28 crc kubenswrapper[4750]: I0220 00:10:28.559654 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:28 crc kubenswrapper[4750]: E0220 00:10:28.560242 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:28 crc kubenswrapper[4750]: E0220 00:10:28.560402 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:28 crc kubenswrapper[4750]: E0220 00:10:28.561225 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:29 crc kubenswrapper[4750]: I0220 00:10:29.559451 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:29 crc kubenswrapper[4750]: E0220 00:10:29.559582 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:29 crc kubenswrapper[4750]: I0220 00:10:29.560875 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:10:29 crc kubenswrapper[4750]: E0220 00:10:29.561184 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hkqgt_openshift-ovn-kubernetes(17ae630a-7ade-407c-b93c-e4bae6d0f8e3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" Feb 20 00:10:30 crc kubenswrapper[4750]: I0220 00:10:30.559877 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:30 crc kubenswrapper[4750]: I0220 00:10:30.560005 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:30 crc kubenswrapper[4750]: I0220 00:10:30.559918 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:30 crc kubenswrapper[4750]: E0220 00:10:30.560105 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:30 crc kubenswrapper[4750]: E0220 00:10:30.560211 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:30 crc kubenswrapper[4750]: E0220 00:10:30.560372 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:31 crc kubenswrapper[4750]: I0220 00:10:31.559003 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:31 crc kubenswrapper[4750]: E0220 00:10:31.559225 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:32 crc kubenswrapper[4750]: I0220 00:10:32.559993 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:32 crc kubenswrapper[4750]: I0220 00:10:32.560084 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:32 crc kubenswrapper[4750]: I0220 00:10:32.560009 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:32 crc kubenswrapper[4750]: E0220 00:10:32.560292 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:32 crc kubenswrapper[4750]: E0220 00:10:32.560451 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:32 crc kubenswrapper[4750]: E0220 00:10:32.560584 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:33 crc kubenswrapper[4750]: I0220 00:10:33.559856 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:33 crc kubenswrapper[4750]: E0220 00:10:33.560047 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:34 crc kubenswrapper[4750]: I0220 00:10:34.559734 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:34 crc kubenswrapper[4750]: E0220 00:10:34.560332 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:34 crc kubenswrapper[4750]: I0220 00:10:34.559786 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:34 crc kubenswrapper[4750]: I0220 00:10:34.559758 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:34 crc kubenswrapper[4750]: E0220 00:10:34.561413 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:34 crc kubenswrapper[4750]: E0220 00:10:34.561535 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:35 crc kubenswrapper[4750]: I0220 00:10:35.558956 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:35 crc kubenswrapper[4750]: E0220 00:10:35.559230 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.264322 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/1.log" Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.266272 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/0.log" Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.266362 4750 generic.go:334] "Generic (PLEG): container finished" podID="1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030" containerID="8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002" exitCode=1 Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.266442 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerDied","Data":"8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002"} Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.266529 4750 scope.go:117] "RemoveContainer" containerID="bf0a384b9732a189a71bba83b7dea1309b579e28e9b9d9d673f36b43e604b7b6" Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.267699 4750 scope.go:117] "RemoveContainer" containerID="8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002" Feb 20 00:10:36 crc kubenswrapper[4750]: E0220 00:10:36.268262 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-dm8lb_openshift-multus(1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030)\"" pod="openshift-multus/multus-dm8lb" podUID="1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030" Feb 20 00:10:36 crc kubenswrapper[4750]: E0220 00:10:36.524693 4750 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.558864 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.558990 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:36 crc kubenswrapper[4750]: E0220 00:10:36.559941 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:36 crc kubenswrapper[4750]: I0220 00:10:36.560094 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:36 crc kubenswrapper[4750]: E0220 00:10:36.560308 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:36 crc kubenswrapper[4750]: E0220 00:10:36.560447 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:36 crc kubenswrapper[4750]: E0220 00:10:36.656365 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 00:10:37 crc kubenswrapper[4750]: I0220 00:10:37.272279 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/1.log" Feb 20 00:10:37 crc kubenswrapper[4750]: I0220 00:10:37.559298 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:37 crc kubenswrapper[4750]: E0220 00:10:37.559826 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:38 crc kubenswrapper[4750]: I0220 00:10:38.559927 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:38 crc kubenswrapper[4750]: I0220 00:10:38.560085 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:38 crc kubenswrapper[4750]: E0220 00:10:38.560224 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:38 crc kubenswrapper[4750]: I0220 00:10:38.560829 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:38 crc kubenswrapper[4750]: E0220 00:10:38.560974 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:38 crc kubenswrapper[4750]: E0220 00:10:38.560819 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:39 crc kubenswrapper[4750]: I0220 00:10:39.559734 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:39 crc kubenswrapper[4750]: E0220 00:10:39.560228 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:40 crc kubenswrapper[4750]: I0220 00:10:40.559297 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:40 crc kubenswrapper[4750]: E0220 00:10:40.559545 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:40 crc kubenswrapper[4750]: I0220 00:10:40.559970 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:40 crc kubenswrapper[4750]: I0220 00:10:40.560062 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:40 crc kubenswrapper[4750]: E0220 00:10:40.560403 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:40 crc kubenswrapper[4750]: E0220 00:10:40.560815 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:41 crc kubenswrapper[4750]: I0220 00:10:41.559081 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:41 crc kubenswrapper[4750]: E0220 00:10:41.559313 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:41 crc kubenswrapper[4750]: E0220 00:10:41.657903 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 00:10:42 crc kubenswrapper[4750]: I0220 00:10:42.559590 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:42 crc kubenswrapper[4750]: I0220 00:10:42.559624 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:42 crc kubenswrapper[4750]: E0220 00:10:42.559809 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:42 crc kubenswrapper[4750]: E0220 00:10:42.559954 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:42 crc kubenswrapper[4750]: I0220 00:10:42.560297 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:42 crc kubenswrapper[4750]: E0220 00:10:42.560447 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:43 crc kubenswrapper[4750]: I0220 00:10:43.558975 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:43 crc kubenswrapper[4750]: E0220 00:10:43.559089 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:44 crc kubenswrapper[4750]: I0220 00:10:44.559896 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:44 crc kubenswrapper[4750]: E0220 00:10:44.560080 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:44 crc kubenswrapper[4750]: I0220 00:10:44.560219 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:44 crc kubenswrapper[4750]: I0220 00:10:44.560333 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:44 crc kubenswrapper[4750]: E0220 00:10:44.561249 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:44 crc kubenswrapper[4750]: E0220 00:10:44.561315 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:44 crc kubenswrapper[4750]: I0220 00:10:44.561766 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:10:45 crc kubenswrapper[4750]: I0220 00:10:45.305458 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/3.log" Feb 20 00:10:45 crc kubenswrapper[4750]: I0220 00:10:45.309933 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerStarted","Data":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} Feb 20 00:10:45 crc kubenswrapper[4750]: I0220 00:10:45.310628 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:10:45 crc kubenswrapper[4750]: I0220 00:10:45.559866 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:45 crc kubenswrapper[4750]: E0220 00:10:45.560093 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:45 crc kubenswrapper[4750]: I0220 00:10:45.566744 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podStartSLOduration=103.566714969 podStartE2EDuration="1m43.566714969s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:10:45.35955938 +0000 UTC m=+129.554395489" watchObservedRunningTime="2026-02-20 00:10:45.566714969 +0000 UTC m=+129.761551048" Feb 20 00:10:45 crc kubenswrapper[4750]: I0220 00:10:45.568577 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tcgh4"] Feb 20 00:10:45 crc kubenswrapper[4750]: I0220 00:10:45.568766 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:45 crc kubenswrapper[4750]: E0220 00:10:45.568987 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:46 crc kubenswrapper[4750]: I0220 00:10:46.559520 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:46 crc kubenswrapper[4750]: I0220 00:10:46.559842 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:46 crc kubenswrapper[4750]: E0220 00:10:46.561791 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:46 crc kubenswrapper[4750]: E0220 00:10:46.561965 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:46 crc kubenswrapper[4750]: E0220 00:10:46.658493 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 00:10:47 crc kubenswrapper[4750]: I0220 00:10:47.558912 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:47 crc kubenswrapper[4750]: I0220 00:10:47.558982 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:47 crc kubenswrapper[4750]: E0220 00:10:47.559035 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:47 crc kubenswrapper[4750]: E0220 00:10:47.559158 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:48 crc kubenswrapper[4750]: I0220 00:10:48.559801 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:48 crc kubenswrapper[4750]: I0220 00:10:48.559816 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:48 crc kubenswrapper[4750]: E0220 00:10:48.559984 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:48 crc kubenswrapper[4750]: E0220 00:10:48.560358 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:49 crc kubenswrapper[4750]: I0220 00:10:49.559232 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:49 crc kubenswrapper[4750]: I0220 00:10:49.559243 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:49 crc kubenswrapper[4750]: E0220 00:10:49.559992 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:49 crc kubenswrapper[4750]: E0220 00:10:49.561637 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:50 crc kubenswrapper[4750]: I0220 00:10:50.559515 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:50 crc kubenswrapper[4750]: E0220 00:10:50.559689 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:50 crc kubenswrapper[4750]: I0220 00:10:50.559929 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:50 crc kubenswrapper[4750]: E0220 00:10:50.560019 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:51 crc kubenswrapper[4750]: I0220 00:10:51.559056 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:51 crc kubenswrapper[4750]: I0220 00:10:51.559109 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:51 crc kubenswrapper[4750]: E0220 00:10:51.559511 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:51 crc kubenswrapper[4750]: E0220 00:10:51.559829 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:51 crc kubenswrapper[4750]: I0220 00:10:51.560037 4750 scope.go:117] "RemoveContainer" containerID="8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002" Feb 20 00:10:51 crc kubenswrapper[4750]: E0220 00:10:51.659732 4750 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 00:10:52 crc kubenswrapper[4750]: I0220 00:10:52.333748 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/1.log" Feb 20 00:10:52 crc kubenswrapper[4750]: I0220 00:10:52.333796 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerStarted","Data":"2edf2c4409f090699cacb0b51a9257cd96eccff1387502cefc49f5308693f05d"} Feb 20 00:10:52 crc kubenswrapper[4750]: I0220 00:10:52.559555 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:52 crc kubenswrapper[4750]: I0220 00:10:52.559728 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:52 crc kubenswrapper[4750]: E0220 00:10:52.561162 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:52 crc kubenswrapper[4750]: E0220 00:10:52.561345 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:53 crc kubenswrapper[4750]: I0220 00:10:53.559776 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:53 crc kubenswrapper[4750]: I0220 00:10:53.559827 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:53 crc kubenswrapper[4750]: E0220 00:10:53.559954 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:53 crc kubenswrapper[4750]: E0220 00:10:53.560099 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:54 crc kubenswrapper[4750]: I0220 00:10:54.559260 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:54 crc kubenswrapper[4750]: I0220 00:10:54.559321 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:54 crc kubenswrapper[4750]: E0220 00:10:54.559805 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:54 crc kubenswrapper[4750]: E0220 00:10:54.559697 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:55 crc kubenswrapper[4750]: I0220 00:10:55.558841 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:55 crc kubenswrapper[4750]: I0220 00:10:55.558867 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:55 crc kubenswrapper[4750]: E0220 00:10:55.558986 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tcgh4" podUID="fac6ee29-99b8-42e7-844d-30e68cbecad3" Feb 20 00:10:55 crc kubenswrapper[4750]: E0220 00:10:55.559052 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 00:10:56 crc kubenswrapper[4750]: I0220 00:10:56.559599 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:56 crc kubenswrapper[4750]: I0220 00:10:56.559660 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:56 crc kubenswrapper[4750]: E0220 00:10:56.563066 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 00:10:56 crc kubenswrapper[4750]: E0220 00:10:56.563229 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 00:10:57 crc kubenswrapper[4750]: I0220 00:10:57.558938 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:10:57 crc kubenswrapper[4750]: I0220 00:10:57.558980 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:10:57 crc kubenswrapper[4750]: I0220 00:10:57.561773 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 00:10:57 crc kubenswrapper[4750]: I0220 00:10:57.561959 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 00:10:57 crc kubenswrapper[4750]: I0220 00:10:57.562180 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 00:10:57 crc kubenswrapper[4750]: I0220 00:10:57.562315 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 00:10:58 crc kubenswrapper[4750]: I0220 00:10:58.559803 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:10:58 crc kubenswrapper[4750]: I0220 00:10:58.559829 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:10:58 crc kubenswrapper[4750]: I0220 00:10:58.563489 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 00:10:58 crc kubenswrapper[4750]: I0220 00:10:58.564227 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 00:11:00 crc kubenswrapper[4750]: I0220 00:11:00.341644 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.418985 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.419059 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.419108 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.419220 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.420223 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.425022 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.425919 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.426847 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.494321 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.520351 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:04 crc kubenswrapper[4750]: E0220 00:11:04.520545 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:13:06.520516563 +0000 UTC m=+270.715352602 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.586084 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:11:04 crc kubenswrapper[4750]: I0220 00:11:04.599639 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 00:11:04 crc kubenswrapper[4750]: W0220 00:11:04.843925 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-e09055d7b651bca9969363c30244edb7f8276c3c141827d0201338a9de4e5000 WatchSource:0}: Error finding container e09055d7b651bca9969363c30244edb7f8276c3c141827d0201338a9de4e5000: Status 404 returned error can't find the container with id e09055d7b651bca9969363c30244edb7f8276c3c141827d0201338a9de4e5000 Feb 20 00:11:04 crc kubenswrapper[4750]: W0220 00:11:04.937460 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-6d01eea0bbc97f3c0e79e4c37cecff1df411453cd5bcc6218d8745369c70baef WatchSource:0}: Error finding container 6d01eea0bbc97f3c0e79e4c37cecff1df411453cd5bcc6218d8745369c70baef: Status 404 returned error can't find the container with id 6d01eea0bbc97f3c0e79e4c37cecff1df411453cd5bcc6218d8745369c70baef Feb 20 00:11:05 crc kubenswrapper[4750]: I0220 00:11:05.385740 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b7db191be69a3fe06e31a53d82c3f7bdd8fc2fe2da8b3f31365b2d060041d2a3"} Feb 20 00:11:05 crc kubenswrapper[4750]: I0220 00:11:05.385812 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e09055d7b651bca9969363c30244edb7f8276c3c141827d0201338a9de4e5000"} Feb 20 00:11:05 crc kubenswrapper[4750]: I0220 00:11:05.388583 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0f3ac46eb7e393ce45bfc763542ad16f859fa1c499aa128c746ecfa0bdd086ad"} Feb 20 00:11:05 crc kubenswrapper[4750]: I0220 00:11:05.388660 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cb415fd81ced741c8fcacdeed5b1f5fdbf61426de765564de8bb0c486abd37bc"} Feb 20 00:11:05 crc kubenswrapper[4750]: I0220 00:11:05.388928 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:11:05 crc kubenswrapper[4750]: I0220 00:11:05.391899 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d21e255650af6ddd2599929b0169d608044d8a0b9426f69c66422fd1912a4239"} Feb 20 00:11:05 crc kubenswrapper[4750]: I0220 00:11:05.391973 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6d01eea0bbc97f3c0e79e4c37cecff1df411453cd5bcc6218d8745369c70baef"} Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.673930 4750 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.729004 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jlgbs"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.729826 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.732942 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gk5dc"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.736197 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxn2h"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.736351 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.739194 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.753184 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.753645 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.753855 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.754019 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.754261 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9bxjm"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.754493 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.754593 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.754716 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.755249 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.755334 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.755737 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.756710 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29525760-6k6m7"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757066 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757210 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757451 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nd2j6"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757657 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-serving-cert\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757694 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-image-import-ca\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757721 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757745 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b4de841e-a559-4470-afe3-8173d516d899-audit-dir\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757766 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-etcd-client\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757791 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-etcd-serving-ca\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757814 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-encryption-config\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757840 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-audit\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757863 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-config\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757888 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b4de841e-a559-4470-afe3-8173d516d899-node-pullsecrets\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.757909 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csqcc\" (UniqueName: \"kubernetes.io/projected/b4de841e-a559-4470-afe3-8173d516d899-kube-api-access-csqcc\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.758196 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.759745 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.759936 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.760085 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.760261 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.764342 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.765104 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.765576 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.766259 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.768264 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5r77d"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.768682 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.769018 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.769291 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.769840 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.770403 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.772850 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jqhqr"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.773255 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.783329 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kw26r"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.784020 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.804755 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.805389 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.815823 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.824484 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.829616 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.830034 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.837596 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.837892 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.838334 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.845068 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.845261 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.849390 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.850338 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.850528 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.850640 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-wl4kz"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.851017 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wl4kz" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.856089 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.856364 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.856519 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857017 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857214 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857278 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857456 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857566 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857614 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857693 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857760 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.857788 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858303 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8vgt\" (UniqueName: \"kubernetes.io/projected/3fc4787c-ec22-4961-be53-906fdd27c5f2-kube-api-access-m8vgt\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858338 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-trusted-ca-bundle\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858369 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858391 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b4de841e-a559-4470-afe3-8173d516d899-audit-dir\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858417 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-etcd-serving-ca\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858440 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-service-ca\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858463 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fkr5\" (UniqueName: \"kubernetes.io/projected/d2450fd9-6aae-4645-bcc7-94a67c7e0fef-kube-api-access-5fkr5\") pod \"cluster-samples-operator-665b6dd947-8mkww\" (UID: \"d2450fd9-6aae-4645-bcc7-94a67c7e0fef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858490 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2blpk\" (UniqueName: \"kubernetes.io/projected/9c6bdc15-60ff-445a-9da6-fa6c926960f3-kube-api-access-2blpk\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858517 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-config\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858540 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c6bdc15-60ff-445a-9da6-fa6c926960f3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs9sq\" (UniqueName: \"kubernetes.io/projected/bcd2fc23-8170-4bb9-90a2-5909cbea778a-kube-api-access-hs9sq\") pod \"image-pruner-29525760-6k6m7\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858585 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-config\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858609 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-oauth-config\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858628 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/347720b4-0171-4b11-9b37-940a17978ee1-config\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858651 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/347720b4-0171-4b11-9b37-940a17978ee1-images\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858677 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-image-import-ca\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858700 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-etcd-client\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858722 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-encryption-config\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858745 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-serving-cert\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858770 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-audit\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858792 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-oauth-serving-cert\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858829 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c6bdc15-60ff-445a-9da6-fa6c926960f3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858855 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d2450fd9-6aae-4645-bcc7-94a67c7e0fef-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8mkww\" (UID: \"d2450fd9-6aae-4645-bcc7-94a67c7e0fef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858889 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b4de841e-a559-4470-afe3-8173d516d899-node-pullsecrets\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858912 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csqcc\" (UniqueName: \"kubernetes.io/projected/b4de841e-a559-4470-afe3-8173d516d899-kube-api-access-csqcc\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858939 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bxvq\" (UniqueName: \"kubernetes.io/projected/347720b4-0171-4b11-9b37-940a17978ee1-kube-api-access-2bxvq\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858961 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bcd2fc23-8170-4bb9-90a2-5909cbea778a-serviceca\") pod \"image-pruner-29525760-6k6m7\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.858987 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-serving-cert\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.859009 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/347720b4-0171-4b11-9b37-940a17978ee1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.859175 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b4de841e-a559-4470-afe3-8173d516d899-audit-dir\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.859661 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.859854 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.859880 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-etcd-serving-ca\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860076 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860262 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860271 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860439 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-config\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860461 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860537 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860634 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860652 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860716 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860744 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.860781 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861101 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861269 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861366 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861437 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-image-import-ca\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861448 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861487 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861584 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.861657 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b4de841e-a559-4470-afe3-8173d516d899-node-pullsecrets\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.862075 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-audit\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.868560 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-encryption-config\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.869739 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-serving-cert\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.878386 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b4de841e-a559-4470-afe3-8173d516d899-etcd-client\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.878979 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.878984 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879357 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879374 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879454 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879497 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879568 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879782 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879803 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879885 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.879904 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880046 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880206 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880234 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880374 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880378 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880516 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880538 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880666 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.880753 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.881486 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.887951 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.888263 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.895653 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.919550 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ddrd9"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.920213 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.920552 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gk5dc"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.920686 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.921199 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.922613 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.922835 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.923208 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.923347 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.923509 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.923642 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.923766 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.923795 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.923941 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.924227 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.924359 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.924460 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.924678 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.924783 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.924859 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.924984 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.925050 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.925913 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.926105 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.926268 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.926405 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.926458 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.926599 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.926784 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.927603 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.933504 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.933613 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.936772 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.936987 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.937740 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.940446 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.942484 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.943006 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.943235 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.943296 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7zf67"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.943773 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.944291 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.944451 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.944487 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.944608 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.947769 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.948319 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vbbs7"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.949399 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4de841e-a559-4470-afe3-8173d516d899-trusted-ca-bundle\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.949752 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.950024 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.950061 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.951071 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.969711 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-trusted-ca-bundle\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.971246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-trusted-ca-bundle\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.982211 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-service-ca\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.982253 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fkr5\" (UniqueName: \"kubernetes.io/projected/d2450fd9-6aae-4645-bcc7-94a67c7e0fef-kube-api-access-5fkr5\") pod \"cluster-samples-operator-665b6dd947-8mkww\" (UID: \"d2450fd9-6aae-4645-bcc7-94a67c7e0fef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.982290 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2blpk\" (UniqueName: \"kubernetes.io/projected/9c6bdc15-60ff-445a-9da6-fa6c926960f3-kube-api-access-2blpk\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.984212 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c6bdc15-60ff-445a-9da6-fa6c926960f3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.984263 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs9sq\" (UniqueName: \"kubernetes.io/projected/bcd2fc23-8170-4bb9-90a2-5909cbea778a-kube-api-access-hs9sq\") pod \"image-pruner-29525760-6k6m7\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.984292 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-config\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.984326 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-oauth-config\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.984354 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/347720b4-0171-4b11-9b37-940a17978ee1-config\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.984378 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/347720b4-0171-4b11-9b37-940a17978ee1-images\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997697 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-serving-cert\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997791 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-oauth-serving-cert\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997823 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c6bdc15-60ff-445a-9da6-fa6c926960f3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997853 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d2450fd9-6aae-4645-bcc7-94a67c7e0fef-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8mkww\" (UID: \"d2450fd9-6aae-4645-bcc7-94a67c7e0fef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997904 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bxvq\" (UniqueName: \"kubernetes.io/projected/347720b4-0171-4b11-9b37-940a17978ee1-kube-api-access-2bxvq\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997932 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bcd2fc23-8170-4bb9-90a2-5909cbea778a-serviceca\") pod \"image-pruner-29525760-6k6m7\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997964 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/347720b4-0171-4b11-9b37-940a17978ee1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.997990 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8vgt\" (UniqueName: \"kubernetes.io/projected/3fc4787c-ec22-4961-be53-906fdd27c5f2-kube-api-access-m8vgt\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.998135 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c6bdc15-60ff-445a-9da6-fa6c926960f3-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.998306 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.998777 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gksdw"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.999142 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bcd2fc23-8170-4bb9-90a2-5909cbea778a-serviceca\") pod \"image-pruner-29525760-6k6m7\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.999168 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-oauth-serving-cert\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.999302 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.999454 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v255n"] Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.999575 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:06 crc kubenswrapper[4750]: I0220 00:11:06.999694 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.000331 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.001171 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-config\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.001908 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-serving-cert\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.002018 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/347720b4-0171-4b11-9b37-940a17978ee1-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.002611 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d2450fd9-6aae-4645-bcc7-94a67c7e0fef-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8mkww\" (UID: \"d2450fd9-6aae-4645-bcc7-94a67c7e0fef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.002748 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.002833 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.003056 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c6bdc15-60ff-445a-9da6-fa6c926960f3-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.003106 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxn2h"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.003246 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3fc4787c-ec22-4961-be53-906fdd27c5f2-service-ca\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.003381 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.003509 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/347720b4-0171-4b11-9b37-940a17978ee1-config\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.003686 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3fc4787c-ec22-4961-be53-906fdd27c5f2-console-oauth-config\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.004160 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jlgbs"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.004824 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.005195 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.005455 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.005845 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csqcc\" (UniqueName: \"kubernetes.io/projected/b4de841e-a559-4470-afe3-8173d516d899-kube-api-access-csqcc\") pod \"apiserver-76f77b778f-jlgbs\" (UID: \"b4de841e-a559-4470-afe3-8173d516d899\") " pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.005959 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.006579 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.006951 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-gtvxw"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.007358 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.007455 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/347720b4-0171-4b11-9b37-940a17978ee1-images\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.007875 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.008520 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.008957 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.009766 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.009888 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.010542 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.011056 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.011600 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.011840 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.012410 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.012979 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.013714 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9bxjm"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.014958 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29525760-6k6m7"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.015904 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.016858 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.017954 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-6f5qm"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.018370 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.021795 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nd2j6"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.021836 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jqhqr"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.021847 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5r77d"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.025476 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.025574 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ddrd9"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.028199 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wl4kz"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.028234 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.029367 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vbbs7"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.030879 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.031923 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.032028 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.033040 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.034038 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4mg6w"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.034638 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.035093 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.036008 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.038239 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.039509 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.041318 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gksdw"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.042401 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7zf67"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.043350 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-8496w"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.044041 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8496w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.045029 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.046032 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.047164 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.048156 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v255n"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.049178 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.050111 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6f5qm"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.051232 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kw26r"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.051520 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.052361 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.054181 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.055104 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.056091 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.057150 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8496w"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.062235 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.068207 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.072359 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.072771 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zs972"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.073962 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.074373 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zs972"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.075445 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.080785 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.092373 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.111565 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.133926 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.152378 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.172735 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.191827 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.212286 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.219331 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-jlgbs"] Feb 20 00:11:07 crc kubenswrapper[4750]: W0220 00:11:07.224217 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4de841e_a559_4470_afe3_8173d516d899.slice/crio-8fb4813ec19c79d4999b839841407331100dc32fcd00de830f1a43c1d474d155 WatchSource:0}: Error finding container 8fb4813ec19c79d4999b839841407331100dc32fcd00de830f1a43c1d474d155: Status 404 returned error can't find the container with id 8fb4813ec19c79d4999b839841407331100dc32fcd00de830f1a43c1d474d155 Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.231235 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.251762 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.272448 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.292033 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.312636 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.332568 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.353229 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.373101 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.392535 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.399565 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" event={"ID":"b4de841e-a559-4470-afe3-8173d516d899","Type":"ContainerStarted","Data":"8fb4813ec19c79d4999b839841407331100dc32fcd00de830f1a43c1d474d155"} Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.412181 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.431175 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.471778 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: E0220 00:11:07.489365 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4de841e_a559_4470_afe3_8173d516d899.slice/crio-8c476290c0d49ebedd3cc39a785449d6e7df3dbae344294d33567f6efbe28b8e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4de841e_a559_4470_afe3_8173d516d899.slice/crio-conmon-8c476290c0d49ebedd3cc39a785449d6e7df3dbae344294d33567f6efbe28b8e.scope\": RecentStats: unable to find data in memory cache]" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.491885 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.511545 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.532833 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.552424 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.572628 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.592408 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.656000 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2blpk\" (UniqueName: \"kubernetes.io/projected/9c6bdc15-60ff-445a-9da6-fa6c926960f3-kube-api-access-2blpk\") pod \"openshift-controller-manager-operator-756b6f6bc6-kfsfb\" (UID: \"9c6bdc15-60ff-445a-9da6-fa6c926960f3\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.679769 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8vgt\" (UniqueName: \"kubernetes.io/projected/3fc4787c-ec22-4961-be53-906fdd27c5f2-kube-api-access-m8vgt\") pod \"console-f9d7485db-kw26r\" (UID: \"3fc4787c-ec22-4961-be53-906fdd27c5f2\") " pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.686384 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.693270 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.698502 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bxvq\" (UniqueName: \"kubernetes.io/projected/347720b4-0171-4b11-9b37-940a17978ee1-kube-api-access-2bxvq\") pod \"machine-api-operator-5694c8668f-gk5dc\" (UID: \"347720b4-0171-4b11-9b37-940a17978ee1\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706216 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-encryption-config\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706303 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f59867a-c7d1-42b8-8648-d50a179e80d0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706333 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5d7c67d4-013f-494c-88dc-079a3accf05b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706352 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-serving-cert\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706370 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4sxx\" (UniqueName: \"kubernetes.io/projected/2678c8d7-202c-48b9-a5ec-cb56277af3a6-kube-api-access-j4sxx\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706469 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-etcd-client\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706546 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-certificates\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706602 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f3fd4f4-5ed4-40de-8632-65184569538c-serving-cert\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706665 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brf8t\" (UniqueName: \"kubernetes.io/projected/1e42c621-3be4-4655-afc9-2a55a4e548b1-kube-api-access-brf8t\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706717 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706768 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5d7c67d4-013f-494c-88dc-079a3accf05b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706813 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-config\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706861 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-tls\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706908 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3fd4f4-5ed4-40de-8632-65184569538c-config\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.706963 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zkmv\" (UniqueName: \"kubernetes.io/projected/ddfac2cb-f987-4870-9453-afd3dce44113-kube-api-access-4zkmv\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707016 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzv29\" (UniqueName: \"kubernetes.io/projected/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-kube-api-access-jzv29\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707073 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707198 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98d80369-7e37-4be7-beea-22dc5ef528c3-serving-cert\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707251 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a972fe8-3fc2-42d8-b697-5addfe880cbe-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707376 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f3fd4f4-5ed4-40de-8632-65184569538c-trusted-ca\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707436 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc572\" (UniqueName: \"kubernetes.io/projected/0f3fd4f4-5ed4-40de-8632-65184569538c-kube-api-access-vc572\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707484 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707539 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs5mp\" (UniqueName: \"kubernetes.io/projected/57cd71b3-f7c5-49c5-9588-cd3107e1371d-kube-api-access-fs5mp\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707592 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707648 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-trusted-ca\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707699 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.707787 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4c40a07-5939-4d83-83af-e22f773ee511-audit-dir\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708002 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-bound-sa-token\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708050 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708102 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708189 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nld4x\" (UniqueName: \"kubernetes.io/projected/98d80369-7e37-4be7-beea-22dc5ef528c3-kube-api-access-nld4x\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708226 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2678c8d7-202c-48b9-a5ec-cb56277af3a6-serving-cert\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708260 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbb82\" (UniqueName: \"kubernetes.io/projected/5f59867a-c7d1-42b8-8648-d50a179e80d0-kube-api-access-rbb82\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708309 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708344 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5f59867a-c7d1-42b8-8648-d50a179e80d0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708380 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57cd71b3-f7c5-49c5-9588-cd3107e1371d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708412 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-serving-cert\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708446 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-config\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708488 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s8kt\" (UniqueName: \"kubernetes.io/projected/8a972fe8-3fc2-42d8-b697-5addfe880cbe-kube-api-access-5s8kt\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708519 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xkkr\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-kube-api-access-9xkkr\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708553 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-audit-policies\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708579 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708638 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwnhs\" (UniqueName: \"kubernetes.io/projected/c4c40a07-5939-4d83-83af-e22f773ee511-kube-api-access-lwnhs\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708674 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708708 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e42c621-3be4-4655-afc9-2a55a4e548b1-auth-proxy-config\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708729 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708752 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f59867a-c7d1-42b8-8648-d50a179e80d0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708785 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708804 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1e42c621-3be4-4655-afc9-2a55a4e548b1-machine-approver-tls\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708825 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-audit-policies\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708843 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.708875 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2678c8d7-202c-48b9-a5ec-cb56277af3a6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709006 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-client-ca\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709025 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709044 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: E0220 00:11:07.709060 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.209041302 +0000 UTC m=+152.403877471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709187 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ddfac2cb-f987-4870-9453-afd3dce44113-audit-dir\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709299 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709343 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e42c621-3be4-4655-afc9-2a55a4e548b1-config\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709367 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57cd71b3-f7c5-49c5-9588-cd3107e1371d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709392 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.709431 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a972fe8-3fc2-42d8-b697-5addfe880cbe-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.710396 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.712603 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.732914 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.753409 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.774584 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.795209 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.810766 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:07 crc kubenswrapper[4750]: E0220 00:11:07.810997 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.310962846 +0000 UTC m=+152.505798935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-audit-policies\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811156 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811200 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s8kt\" (UniqueName: \"kubernetes.io/projected/8a972fe8-3fc2-42d8-b697-5addfe880cbe-kube-api-access-5s8kt\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811239 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/37e0b2a3-df5f-4e4f-8146-ab0c30593448-images\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811271 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37e0b2a3-df5f-4e4f-8146-ab0c30593448-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811305 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp6xr\" (UniqueName: \"kubernetes.io/projected/4d498c49-4df5-4e03-839f-739dafa7086f-kube-api-access-fp6xr\") pod \"multus-admission-controller-857f4d67dd-7zf67\" (UID: \"4d498c49-4df5-4e03-839f-739dafa7086f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811337 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb30ed35-d423-4c83-a8e3-3294db1fd87d-srv-cert\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811374 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwnhs\" (UniqueName: \"kubernetes.io/projected/c4c40a07-5939-4d83-83af-e22f773ee511-kube-api-access-lwnhs\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811405 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-metrics-tls\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811447 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811480 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e42c621-3be4-4655-afc9-2a55a4e548b1-auth-proxy-config\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811563 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811603 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f59867a-c7d1-42b8-8648-d50a179e80d0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811639 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811671 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-audit-policies\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811708 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40b8f502-e114-4817-82dd-fc33d6085119-service-ca-bundle\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811741 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-client-ca\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811798 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d-cert\") pod \"ingress-canary-6f5qm\" (UID: \"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d\") " pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811832 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c602074-9ed8-4385-a338-40fefebb4924-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z57xn\" (UID: \"1c602074-9ed8-4385-a338-40fefebb4924\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811906 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811941 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22sm6\" (UniqueName: \"kubernetes.io/projected/40b8f502-e114-4817-82dd-fc33d6085119-kube-api-access-22sm6\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.811973 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c84179f9-cfa5-4d80-a5ba-67831c067e0d-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812011 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd94t\" (UniqueName: \"kubernetes.io/projected/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-kube-api-access-wd94t\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812049 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57cd71b3-f7c5-49c5-9588-cd3107e1371d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812080 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-srv-cert\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812153 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fef276de-096a-4b26-afdc-5a946f8a9502-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812204 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f59867a-c7d1-42b8-8648-d50a179e80d0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812236 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjbd6\" (UniqueName: \"kubernetes.io/projected/14a5a02a-4af8-4794-af5d-4a12c4098e3a-kube-api-access-vjbd6\") pod \"downloads-7954f5f757-wl4kz\" (UID: \"14a5a02a-4af8-4794-af5d-4a12c4098e3a\") " pod="openshift-console/downloads-7954f5f757-wl4kz" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812271 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812309 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5d7c67d4-013f-494c-88dc-079a3accf05b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812345 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4sxx\" (UniqueName: \"kubernetes.io/projected/2678c8d7-202c-48b9-a5ec-cb56277af3a6-kube-api-access-j4sxx\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812380 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-ca\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812412 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-metrics-certs\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812443 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c38baaf-c316-46a7-a224-9ae49128bc6f-trusted-ca\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812477 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk9sx\" (UniqueName: \"kubernetes.io/projected/7c38baaf-c316-46a7-a224-9ae49128bc6f-kube-api-access-gk9sx\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812514 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-etcd-client\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812571 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2skk\" (UniqueName: \"kubernetes.io/projected/02afee9d-7374-42ac-bef1-a2b321cb6418-kube-api-access-n2skk\") pod \"dns-operator-744455d44c-vbbs7\" (UID: \"02afee9d-7374-42ac-bef1-a2b321cb6418\") " pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812608 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4x9r\" (UniqueName: \"kubernetes.io/projected/37e0b2a3-df5f-4e4f-8146-ab0c30593448-kube-api-access-h4x9r\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812642 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97fa91c6-34c8-454a-af96-84f2b73f513f-serving-cert\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812677 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-config\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812711 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e2d40b9-179b-473d-b639-6c2ba54f2815-serving-cert\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812744 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78vmh\" (UniqueName: \"kubernetes.io/projected/76b558c0-3be2-4d0d-af06-7cec72d518c3-kube-api-access-78vmh\") pod \"migrator-59844c95c7-8qz7f\" (UID: \"76b558c0-3be2-4d0d-af06-7cec72d518c3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812778 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5gqg\" (UniqueName: \"kubernetes.io/projected/8e9e25a5-cefa-401e-9126-562dc4a5c277-kube-api-access-w5gqg\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812810 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-default-certificate\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812859 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-serving-cert\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812895 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzv29\" (UniqueName: \"kubernetes.io/projected/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-kube-api-access-jzv29\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.812929 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lrtn\" (UniqueName: \"kubernetes.io/projected/1e2d40b9-179b-473d-b639-6c2ba54f2815-kube-api-access-6lrtn\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813004 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-proxy-tls\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813040 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce0240f0-b334-484f-8037-040359dda7f7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813071 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd7jv\" (UniqueName: \"kubernetes.io/projected/a7b98640-cfde-41a4-81c1-a44d16db3219-kube-api-access-gd7jv\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813104 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgjzf\" (UniqueName: \"kubernetes.io/projected/97fa91c6-34c8-454a-af96-84f2b73f513f-kube-api-access-bgjzf\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813166 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9e5606c6-cb98-484a-ac5a-746ef795372c-node-bootstrap-token\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813205 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-registration-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813261 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc572\" (UniqueName: \"kubernetes.io/projected/0f3fd4f4-5ed4-40de-8632-65184569538c-kube-api-access-vc572\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813648 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813784 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs5mp\" (UniqueName: \"kubernetes.io/projected/57cd71b3-f7c5-49c5-9588-cd3107e1371d-kube-api-access-fs5mp\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813826 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5acb54a3-8f1a-4d97-8200-1da490906e99-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.813876 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-trusted-ca\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814223 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dml4g\" (UniqueName: \"kubernetes.io/projected/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-kube-api-access-dml4g\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814285 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814331 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tln59\" (UniqueName: \"kubernetes.io/projected/eb30ed35-d423-4c83-a8e3-3294db1fd87d-kube-api-access-tln59\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814371 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814629 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02afee9d-7374-42ac-bef1-a2b321cb6418-metrics-tls\") pod \"dns-operator-744455d44c-vbbs7\" (UID: \"02afee9d-7374-42ac-bef1-a2b321cb6418\") " pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814679 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nld4x\" (UniqueName: \"kubernetes.io/projected/98d80369-7e37-4be7-beea-22dc5ef528c3-kube-api-access-nld4x\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814724 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2678c8d7-202c-48b9-a5ec-cb56277af3a6-serving-cert\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814945 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8e9e25a5-cefa-401e-9126-562dc4a5c277-signing-cabundle\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.814993 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5f59867a-c7d1-42b8-8648-d50a179e80d0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.815042 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1c36fa76-92b7-43f8-9e82-97d2506cfb20-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sz9qr\" (UID: \"1c36fa76-92b7-43f8-9e82-97d2506cfb20\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.815229 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-profile-collector-cert\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: E0220 00:11:07.818659 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.318638258 +0000 UTC m=+152.513474327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.819602 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57cd71b3-f7c5-49c5-9588-cd3107e1371d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.820180 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.820537 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1e42c621-3be4-4655-afc9-2a55a4e548b1-auth-proxy-config\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.820761 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-audit-policies\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.821802 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.822605 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-client-ca\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.824813 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-config\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.825701 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5f59867a-c7d1-42b8-8648-d50a179e80d0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.825845 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.826399 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2678c8d7-202c-48b9-a5ec-cb56277af3a6-serving-cert\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.815459 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9e5606c6-cb98-484a-ac5a-746ef795372c-certs\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.827288 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-secret-volume\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.827351 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-serving-cert\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.827398 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.828865 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-config-volume\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.828926 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-plugins-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.829077 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xkkr\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-kube-api-access-9xkkr\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.829169 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-audit-policies\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.829898 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddfac2cb-f987-4870-9453-afd3dce44113-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.831171 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-trusted-ca\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.831657 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-config-volume\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.831839 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f59867a-c7d1-42b8-8648-d50a179e80d0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.831895 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1e42c621-3be4-4655-afc9-2a55a4e548b1-machine-approver-tls\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.832052 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.832059 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.832155 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2678c8d7-202c-48b9-a5ec-cb56277af3a6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.832622 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-service-ca\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.832665 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6hfr\" (UniqueName: \"kubernetes.io/projected/c84179f9-cfa5-4d80-a5ba-67831c067e0d-kube-api-access-s6hfr\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.833219 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fef276de-096a-4b26-afdc-5a946f8a9502-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.833859 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5d7c67d4-013f-494c-88dc-079a3accf05b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.833981 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-etcd-client\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.834057 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2678c8d7-202c-48b9-a5ec-cb56277af3a6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.834253 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-serving-cert\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.834308 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.834353 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.834399 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-stats-auth\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.834543 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c84179f9-cfa5-4d80-a5ba-67831c067e0d-webhook-cert\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.834821 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ddfac2cb-f987-4870-9453-afd3dce44113-audit-dir\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.835149 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ddfac2cb-f987-4870-9453-afd3dce44113-audit-dir\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.835206 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-mountpoint-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.835417 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-client-ca\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.836404 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e42c621-3be4-4655-afc9-2a55a4e548b1-config\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.836730 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.836874 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a972fe8-3fc2-42d8-b697-5addfe880cbe-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.837576 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e42c621-3be4-4655-afc9-2a55a4e548b1-config\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.837479 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-encryption-config\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.838042 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce0240f0-b334-484f-8037-040359dda7f7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.838165 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-serving-cert\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.838243 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmbdx\" (UniqueName: \"kubernetes.io/projected/1c602074-9ed8-4385-a338-40fefebb4924-kube-api-access-xmbdx\") pod \"control-plane-machine-set-operator-78cbb6b69f-z57xn\" (UID: \"1c602074-9ed8-4385-a338-40fefebb4924\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.838348 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef276de-096a-4b26-afdc-5a946f8a9502-config\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.838985 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839065 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l49cw\" (UniqueName: \"kubernetes.io/projected/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-kube-api-access-l49cw\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839467 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4d498c49-4df5-4e03-839f-739dafa7086f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7zf67\" (UID: \"4d498c49-4df5-4e03-839f-739dafa7086f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839541 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c84179f9-cfa5-4d80-a5ba-67831c067e0d-tmpfs\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839607 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839650 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-certificates\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839716 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f3fd4f4-5ed4-40de-8632-65184569538c-serving-cert\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839766 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brf8t\" (UniqueName: \"kubernetes.io/projected/1e42c621-3be4-4655-afc9-2a55a4e548b1-kube-api-access-brf8t\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839871 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839926 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5acb54a3-8f1a-4d97-8200-1da490906e99-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.839978 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb30ed35-d423-4c83-a8e3-3294db1fd87d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840022 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5d7c67d4-013f-494c-88dc-079a3accf05b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840251 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a972fe8-3fc2-42d8-b697-5addfe880cbe-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840275 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840324 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgjnd\" (UniqueName: \"kubernetes.io/projected/55531783-59c5-4d74-b4af-d9e89c0a8421-kube-api-access-tgjnd\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840505 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37e0b2a3-df5f-4e4f-8146-ab0c30593448-proxy-tls\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840536 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkc4f\" (UniqueName: \"kubernetes.io/projected/1c36fa76-92b7-43f8-9e82-97d2506cfb20-kube-api-access-gkc4f\") pod \"package-server-manager-789f6589d5-sz9qr\" (UID: \"1c36fa76-92b7-43f8-9e82-97d2506cfb20\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840576 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkmbv\" (UniqueName: \"kubernetes.io/projected/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-kube-api-access-mkmbv\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.841783 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ddfac2cb-f987-4870-9453-afd3dce44113-encryption-config\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.848021 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.840611 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-tls\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.850427 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-certificates\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.841932 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.851209 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f3fd4f4-5ed4-40de-8632-65184569538c-serving-cert\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.851861 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-serving-cert\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.851999 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3fd4f4-5ed4-40de-8632-65184569538c-config\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.852032 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8e9e25a5-cefa-401e-9126-562dc4a5c277-signing-key\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.852096 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zkmv\" (UniqueName: \"kubernetes.io/projected/ddfac2cb-f987-4870-9453-afd3dce44113-kube-api-access-4zkmv\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.852213 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.844213 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.852136 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn5bm\" (UniqueName: \"kubernetes.io/projected/9e5606c6-cb98-484a-ac5a-746ef795372c-kube-api-access-kn5bm\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.852419 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3fd4f4-5ed4-40de-8632-65184569538c-config\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.852478 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.853184 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1e42c621-3be4-4655-afc9-2a55a4e548b1-machine-approver-tls\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.853215 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.853362 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5d7c67d4-013f-494c-88dc-079a3accf05b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.853751 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.853827 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98d80369-7e37-4be7-beea-22dc5ef528c3-serving-cert\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.853883 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a972fe8-3fc2-42d8-b697-5addfe880cbe-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.853950 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-socket-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.854030 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97fa91c6-34c8-454a-af96-84f2b73f513f-config\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.854105 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f3fd4f4-5ed4-40de-8632-65184569538c-trusted-ca\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855522 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855661 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5acb54a3-8f1a-4d97-8200-1da490906e99-config\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855677 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f3fd4f4-5ed4-40de-8632-65184569538c-trusted-ca\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855699 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7c38baaf-c316-46a7-a224-9ae49128bc6f-metrics-tls\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855776 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855850 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkb7v\" (UniqueName: \"kubernetes.io/projected/0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d-kube-api-access-xkb7v\") pod \"ingress-canary-6f5qm\" (UID: \"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d\") " pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855882 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hjcc\" (UniqueName: \"kubernetes.io/projected/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-kube-api-access-4hjcc\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855962 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4c40a07-5939-4d83-83af-e22f773ee511-audit-dir\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.855994 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce0240f0-b334-484f-8037-040359dda7f7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.856060 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c38baaf-c316-46a7-a224-9ae49128bc6f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.856174 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-bound-sa-token\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.856264 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.856267 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-csi-data-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.857072 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-tls\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.857369 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4c40a07-5939-4d83-83af-e22f773ee511-audit-dir\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.857846 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-config\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.857904 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbb82\" (UniqueName: \"kubernetes.io/projected/5f59867a-c7d1-42b8-8648-d50a179e80d0-kube-api-access-rbb82\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.857958 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.857992 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57cd71b3-f7c5-49c5-9588-cd3107e1371d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.858024 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-config\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.858057 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-client\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.858071 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.858089 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-config\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.858595 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.859008 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-config\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.859430 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.860594 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a972fe8-3fc2-42d8-b697-5addfe880cbe-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.860754 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.861486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57cd71b3-f7c5-49c5-9588-cd3107e1371d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.861684 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98d80369-7e37-4be7-beea-22dc5ef528c3-serving-cert\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.865174 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.873457 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.907840 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kw26r"] Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.913555 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.917737 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs9sq\" (UniqueName: \"kubernetes.io/projected/bcd2fc23-8170-4bb9-90a2-5909cbea778a-kube-api-access-hs9sq\") pod \"image-pruner-29525760-6k6m7\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.932469 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.941554 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.951584 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.951827 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gk5dc"] Feb 20 00:11:07 crc kubenswrapper[4750]: W0220 00:11:07.955367 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod347720b4_0171_4b11_9b37_940a17978ee1.slice/crio-6602bffdb5ab9ba602f801dec5c2e9cbb7b3b00cb10c7aa03c8e8252120fd420 WatchSource:0}: Error finding container 6602bffdb5ab9ba602f801dec5c2e9cbb7b3b00cb10c7aa03c8e8252120fd420: Status 404 returned error can't find the container with id 6602bffdb5ab9ba602f801dec5c2e9cbb7b3b00cb10c7aa03c8e8252120fd420 Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959028 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:07 crc kubenswrapper[4750]: E0220 00:11:07.959235 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.459202465 +0000 UTC m=+152.654038524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959320 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959346 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40b8f502-e114-4817-82dd-fc33d6085119-service-ca-bundle\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959749 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c602074-9ed8-4385-a338-40fefebb4924-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z57xn\" (UID: \"1c602074-9ed8-4385-a338-40fefebb4924\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959798 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d-cert\") pod \"ingress-canary-6f5qm\" (UID: \"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d\") " pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959834 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22sm6\" (UniqueName: \"kubernetes.io/projected/40b8f502-e114-4817-82dd-fc33d6085119-kube-api-access-22sm6\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959858 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c84179f9-cfa5-4d80-a5ba-67831c067e0d-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959881 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd94t\" (UniqueName: \"kubernetes.io/projected/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-kube-api-access-wd94t\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959917 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fef276de-096a-4b26-afdc-5a946f8a9502-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959941 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-srv-cert\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959963 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjbd6\" (UniqueName: \"kubernetes.io/projected/14a5a02a-4af8-4794-af5d-4a12c4098e3a-kube-api-access-vjbd6\") pod \"downloads-7954f5f757-wl4kz\" (UID: \"14a5a02a-4af8-4794-af5d-4a12c4098e3a\") " pod="openshift-console/downloads-7954f5f757-wl4kz" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.959985 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960020 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-ca\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960050 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-metrics-certs\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960077 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c38baaf-c316-46a7-a224-9ae49128bc6f-trusted-ca\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960100 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk9sx\" (UniqueName: \"kubernetes.io/projected/7c38baaf-c316-46a7-a224-9ae49128bc6f-kube-api-access-gk9sx\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960173 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2skk\" (UniqueName: \"kubernetes.io/projected/02afee9d-7374-42ac-bef1-a2b321cb6418-kube-api-access-n2skk\") pod \"dns-operator-744455d44c-vbbs7\" (UID: \"02afee9d-7374-42ac-bef1-a2b321cb6418\") " pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960196 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4x9r\" (UniqueName: \"kubernetes.io/projected/37e0b2a3-df5f-4e4f-8146-ab0c30593448-kube-api-access-h4x9r\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960218 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97fa91c6-34c8-454a-af96-84f2b73f513f-serving-cert\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960244 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5gqg\" (UniqueName: \"kubernetes.io/projected/8e9e25a5-cefa-401e-9126-562dc4a5c277-kube-api-access-w5gqg\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960267 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-default-certificate\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: E0220 00:11:07.960286 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.460268075 +0000 UTC m=+152.655104144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960369 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e2d40b9-179b-473d-b639-6c2ba54f2815-serving-cert\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960429 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78vmh\" (UniqueName: \"kubernetes.io/projected/76b558c0-3be2-4d0d-af06-7cec72d518c3-kube-api-access-78vmh\") pod \"migrator-59844c95c7-8qz7f\" (UID: \"76b558c0-3be2-4d0d-af06-7cec72d518c3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960481 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-serving-cert\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960506 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lrtn\" (UniqueName: \"kubernetes.io/projected/1e2d40b9-179b-473d-b639-6c2ba54f2815-kube-api-access-6lrtn\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960564 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-proxy-tls\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960624 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd7jv\" (UniqueName: \"kubernetes.io/projected/a7b98640-cfde-41a4-81c1-a44d16db3219-kube-api-access-gd7jv\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960650 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgjzf\" (UniqueName: \"kubernetes.io/projected/97fa91c6-34c8-454a-af96-84f2b73f513f-kube-api-access-bgjzf\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960673 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce0240f0-b334-484f-8037-040359dda7f7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960695 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-registration-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960720 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9e5606c6-cb98-484a-ac5a-746ef795372c-node-bootstrap-token\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960758 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5acb54a3-8f1a-4d97-8200-1da490906e99-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960781 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dml4g\" (UniqueName: \"kubernetes.io/projected/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-kube-api-access-dml4g\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960805 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tln59\" (UniqueName: \"kubernetes.io/projected/eb30ed35-d423-4c83-a8e3-3294db1fd87d-kube-api-access-tln59\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960827 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02afee9d-7374-42ac-bef1-a2b321cb6418-metrics-tls\") pod \"dns-operator-744455d44c-vbbs7\" (UID: \"02afee9d-7374-42ac-bef1-a2b321cb6418\") " pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960848 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8e9e25a5-cefa-401e-9126-562dc4a5c277-signing-cabundle\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960878 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-profile-collector-cert\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960899 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9e5606c6-cb98-484a-ac5a-746ef795372c-certs\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960929 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1c36fa76-92b7-43f8-9e82-97d2506cfb20-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sz9qr\" (UID: \"1c36fa76-92b7-43f8-9e82-97d2506cfb20\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960958 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-secret-volume\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.960988 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961055 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-config-volume\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961087 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-plugins-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961140 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-config-volume\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961167 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-service-ca\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961189 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6hfr\" (UniqueName: \"kubernetes.io/projected/c84179f9-cfa5-4d80-a5ba-67831c067e0d-kube-api-access-s6hfr\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961211 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fef276de-096a-4b26-afdc-5a946f8a9502-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961236 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-stats-auth\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961257 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c84179f9-cfa5-4d80-a5ba-67831c067e0d-webhook-cert\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961281 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-mountpoint-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961304 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-client-ca\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961320 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961333 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce0240f0-b334-484f-8037-040359dda7f7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961367 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmbdx\" (UniqueName: \"kubernetes.io/projected/1c602074-9ed8-4385-a338-40fefebb4924-kube-api-access-xmbdx\") pod \"control-plane-machine-set-operator-78cbb6b69f-z57xn\" (UID: \"1c602074-9ed8-4385-a338-40fefebb4924\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961388 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef276de-096a-4b26-afdc-5a946f8a9502-config\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961409 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l49cw\" (UniqueName: \"kubernetes.io/projected/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-kube-api-access-l49cw\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961426 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4d498c49-4df5-4e03-839f-739dafa7086f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7zf67\" (UID: \"4d498c49-4df5-4e03-839f-739dafa7086f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961444 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c84179f9-cfa5-4d80-a5ba-67831c067e0d-tmpfs\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961469 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5acb54a3-8f1a-4d97-8200-1da490906e99-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961512 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb30ed35-d423-4c83-a8e3-3294db1fd87d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-registration-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961684 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkc4f\" (UniqueName: \"kubernetes.io/projected/1c36fa76-92b7-43f8-9e82-97d2506cfb20-kube-api-access-gkc4f\") pod \"package-server-manager-789f6589d5-sz9qr\" (UID: \"1c36fa76-92b7-43f8-9e82-97d2506cfb20\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961706 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkmbv\" (UniqueName: \"kubernetes.io/projected/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-kube-api-access-mkmbv\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961726 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgjnd\" (UniqueName: \"kubernetes.io/projected/55531783-59c5-4d74-b4af-d9e89c0a8421-kube-api-access-tgjnd\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961742 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37e0b2a3-df5f-4e4f-8146-ab0c30593448-proxy-tls\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961740 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-plugins-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961762 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8e9e25a5-cefa-401e-9126-562dc4a5c277-signing-key\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961780 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn5bm\" (UniqueName: \"kubernetes.io/projected/9e5606c6-cb98-484a-ac5a-746ef795372c-kube-api-access-kn5bm\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961804 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961822 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-socket-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961839 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97fa91c6-34c8-454a-af96-84f2b73f513f-config\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961864 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5acb54a3-8f1a-4d97-8200-1da490906e99-config\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961880 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7c38baaf-c316-46a7-a224-9ae49128bc6f-metrics-tls\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961926 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkb7v\" (UniqueName: \"kubernetes.io/projected/0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d-kube-api-access-xkb7v\") pod \"ingress-canary-6f5qm\" (UID: \"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d\") " pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961942 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hjcc\" (UniqueName: \"kubernetes.io/projected/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-kube-api-access-4hjcc\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961959 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce0240f0-b334-484f-8037-040359dda7f7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961974 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c38baaf-c316-46a7-a224-9ae49128bc6f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.961997 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-csi-data-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962017 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-config\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962040 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-client\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962041 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-ca\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962059 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-config\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962150 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/37e0b2a3-df5f-4e4f-8146-ab0c30593448-images\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962197 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37e0b2a3-df5f-4e4f-8146-ab0c30593448-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962239 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp6xr\" (UniqueName: \"kubernetes.io/projected/4d498c49-4df5-4e03-839f-739dafa7086f-kube-api-access-fp6xr\") pod \"multus-admission-controller-857f4d67dd-7zf67\" (UID: \"4d498c49-4df5-4e03-839f-739dafa7086f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962274 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb30ed35-d423-4c83-a8e3-3294db1fd87d-srv-cert\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962318 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-metrics-tls\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962477 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-config\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.962773 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-config-volume\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.963092 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef276de-096a-4b26-afdc-5a946f8a9502-config\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.963204 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-service-ca\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.963487 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/37e0b2a3-df5f-4e4f-8146-ab0c30593448-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.963523 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-socket-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.963735 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-mountpoint-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.964507 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-serving-cert\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.965247 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c84179f9-cfa5-4d80-a5ba-67831c067e0d-tmpfs\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.965256 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5acb54a3-8f1a-4d97-8200-1da490906e99-config\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.965763 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-proxy-tls\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.966025 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-client-ca\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.966338 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-srv-cert\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.968741 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fef276de-096a-4b26-afdc-5a946f8a9502-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.969329 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a7b98640-cfde-41a4-81c1-a44d16db3219-csi-data-dir\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.969505 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5acb54a3-8f1a-4d97-8200-1da490906e99-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.970913 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/eb30ed35-d423-4c83-a8e3-3294db1fd87d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.971372 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.971917 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4d498c49-4df5-4e03-839f-739dafa7086f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7zf67\" (UID: \"4d498c49-4df5-4e03-839f-739dafa7086f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.973480 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/eb30ed35-d423-4c83-a8e3-3294db1fd87d-srv-cert\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.973620 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.973486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.974421 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e2d40b9-179b-473d-b639-6c2ba54f2815-serving-cert\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.974901 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1c36fa76-92b7-43f8-9e82-97d2506cfb20-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-sz9qr\" (UID: \"1c36fa76-92b7-43f8-9e82-97d2506cfb20\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.975959 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-config\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.975984 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-secret-volume\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.976400 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-etcd-client\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.976830 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-profile-collector-cert\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.981566 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8e9e25a5-cefa-401e-9126-562dc4a5c277-signing-key\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:07 crc kubenswrapper[4750]: I0220 00:11:07.992182 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.003469 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8e9e25a5-cefa-401e-9126-562dc4a5c277-signing-cabundle\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.010421 4750 request.go:700] Waited for 1.007876444s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.012065 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.031581 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.055828 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.063337 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.064147 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.564106513 +0000 UTC m=+152.758942572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.066656 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/02afee9d-7374-42ac-bef1-a2b321cb6418-metrics-tls\") pod \"dns-operator-744455d44c-vbbs7\" (UID: \"02afee9d-7374-42ac-bef1-a2b321cb6418\") " pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.072187 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.108959 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fkr5\" (UniqueName: \"kubernetes.io/projected/d2450fd9-6aae-4645-bcc7-94a67c7e0fef-kube-api-access-5fkr5\") pod \"cluster-samples-operator-665b6dd947-8mkww\" (UID: \"d2450fd9-6aae-4645-bcc7-94a67c7e0fef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.115014 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.132619 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.134702 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb"] Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.144862 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c602074-9ed8-4385-a338-40fefebb4924-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z57xn\" (UID: \"1c602074-9ed8-4385-a338-40fefebb4924\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:08 crc kubenswrapper[4750]: W0220 00:11:08.147399 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c6bdc15_60ff_445a_9da6_fa6c926960f3.slice/crio-750fcbe22075afcd8c3dbd2b685c2632b689bae417f7cb741970f66bf22dc5d9 WatchSource:0}: Error finding container 750fcbe22075afcd8c3dbd2b685c2632b689bae417f7cb741970f66bf22dc5d9: Status 404 returned error can't find the container with id 750fcbe22075afcd8c3dbd2b685c2632b689bae417f7cb741970f66bf22dc5d9 Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.151034 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.154404 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.165853 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.166310 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.666289804 +0000 UTC m=+152.861125863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.172522 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.178820 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7c38baaf-c316-46a7-a224-9ae49128bc6f-metrics-tls\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.198499 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.203794 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7c38baaf-c316-46a7-a224-9ae49128bc6f-trusted-ca\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.214630 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.231731 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.234365 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.251999 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.265667 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-default-certificate\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.266647 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.267062 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.767019636 +0000 UTC m=+152.961855695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.272435 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.275725 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-stats-auth\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.295729 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.310350 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29525760-6k6m7"] Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.311548 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 00:11:08 crc kubenswrapper[4750]: W0220 00:11:08.320200 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcd2fc23_8170_4bb9_90a2_5909cbea778a.slice/crio-65ad14a9a75f0ae998193984f441ec064072e28bdc05228cb6db7d98dc905ae3 WatchSource:0}: Error finding container 65ad14a9a75f0ae998193984f441ec064072e28bdc05228cb6db7d98dc905ae3: Status 404 returned error can't find the container with id 65ad14a9a75f0ae998193984f441ec064072e28bdc05228cb6db7d98dc905ae3 Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.325031 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40b8f502-e114-4817-82dd-fc33d6085119-metrics-certs\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.332575 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.354168 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.361665 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40b8f502-e114-4817-82dd-fc33d6085119-service-ca-bundle\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.368652 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.369092 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.869073525 +0000 UTC m=+153.063909574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.373030 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.392538 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.409986 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" event={"ID":"9c6bdc15-60ff-445a-9da6-fa6c926960f3","Type":"ContainerStarted","Data":"a545aea6add8b642987052b88fe08114d85f9d6af1587e67db614464c3c2eb00"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.410035 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" event={"ID":"9c6bdc15-60ff-445a-9da6-fa6c926960f3","Type":"ContainerStarted","Data":"750fcbe22075afcd8c3dbd2b685c2632b689bae417f7cb741970f66bf22dc5d9"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.413871 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.415977 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww"] Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.416611 4750 generic.go:334] "Generic (PLEG): container finished" podID="b4de841e-a559-4470-afe3-8173d516d899" containerID="8c476290c0d49ebedd3cc39a785449d6e7df3dbae344294d33567f6efbe28b8e" exitCode=0 Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.417259 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" event={"ID":"b4de841e-a559-4470-afe3-8173d516d899","Type":"ContainerDied","Data":"8c476290c0d49ebedd3cc39a785449d6e7df3dbae344294d33567f6efbe28b8e"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.422606 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" event={"ID":"347720b4-0171-4b11-9b37-940a17978ee1","Type":"ContainerStarted","Data":"55e474ce858ba73d3ac72ef304e50d32dfa189c7915e3e68d28787ed054d5fea"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.422635 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" event={"ID":"347720b4-0171-4b11-9b37-940a17978ee1","Type":"ContainerStarted","Data":"7824df50d3c36885bb9ed3980914b81c87085f61d40db95160ec13c73a128705"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.422646 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" event={"ID":"347720b4-0171-4b11-9b37-940a17978ee1","Type":"ContainerStarted","Data":"6602bffdb5ab9ba602f801dec5c2e9cbb7b3b00cb10c7aa03c8e8252120fd420"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.426100 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29525760-6k6m7" event={"ID":"bcd2fc23-8170-4bb9-90a2-5909cbea778a","Type":"ContainerStarted","Data":"65ad14a9a75f0ae998193984f441ec064072e28bdc05228cb6db7d98dc905ae3"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.427760 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kw26r" event={"ID":"3fc4787c-ec22-4961-be53-906fdd27c5f2","Type":"ContainerStarted","Data":"38edba44fc6bd1c1827a51f54c9b368a71df26fc50cd6bfb029379c8c37e51b7"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.427799 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kw26r" event={"ID":"3fc4787c-ec22-4961-be53-906fdd27c5f2","Type":"ContainerStarted","Data":"fb7a5c9da81f3dcc5e528a2d69ed11a5cbd395dc5fd6b54d271483ee7fa2c9fb"} Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.432057 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.451890 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.459566 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c84179f9-cfa5-4d80-a5ba-67831c067e0d-webhook-cert\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.464561 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c84179f9-cfa5-4d80-a5ba-67831c067e0d-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.469353 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.470878 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:08.970858345 +0000 UTC m=+153.165694394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.471820 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.473484 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/37e0b2a3-df5f-4e4f-8146-ab0c30593448-images\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.492643 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.512804 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.519998 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/37e0b2a3-df5f-4e4f-8146-ab0c30593448-proxy-tls\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.531575 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.553165 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.571972 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.572833 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.573320 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.073307084 +0000 UTC m=+153.268143133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.585437 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97fa91c6-34c8-454a-af96-84f2b73f513f-serving-cert\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.591719 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.595616 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97fa91c6-34c8-454a-af96-84f2b73f513f-config\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.611607 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.631427 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.652039 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.672784 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.673970 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.674643 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.174627763 +0000 UTC m=+153.369463812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.685955 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce0240f0-b334-484f-8037-040359dda7f7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.692653 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.702188 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce0240f0-b334-484f-8037-040359dda7f7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.712374 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.732028 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.751974 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.763963 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d-cert\") pod \"ingress-canary-6f5qm\" (UID: \"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d\") " pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.772580 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.775321 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.775812 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.275802287 +0000 UTC m=+153.470638336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.792421 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.806819 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9e5606c6-cb98-484a-ac5a-746ef795372c-node-bootstrap-token\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.811616 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.822107 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9e5606c6-cb98-484a-ac5a-746ef795372c-certs\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.831956 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.852306 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.860430 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-config-volume\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.872105 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.876511 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.876649 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.376631811 +0000 UTC m=+153.571467850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.877143 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.877464 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.377455943 +0000 UTC m=+153.572291992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.893745 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.908902 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-metrics-tls\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.911893 4750 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.931607 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.952485 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.978715 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.978895 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.478872224 +0000 UTC m=+153.673708273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:08 crc kubenswrapper[4750]: I0220 00:11:08.979366 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:08 crc kubenswrapper[4750]: E0220 00:11:08.979744 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.479718418 +0000 UTC m=+153.674554467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.006063 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s8kt\" (UniqueName: \"kubernetes.io/projected/8a972fe8-3fc2-42d8-b697-5addfe880cbe-kube-api-access-5s8kt\") pod \"openshift-apiserver-operator-796bbdcf4f-mn2nx\" (UID: \"8a972fe8-3fc2-42d8-b697-5addfe880cbe\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.010718 4750 request.go:700] Waited for 1.189934884s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console-operator/serviceaccounts/console-operator/token Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.033819 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc572\" (UniqueName: \"kubernetes.io/projected/0f3fd4f4-5ed4-40de-8632-65184569538c-kube-api-access-vc572\") pod \"console-operator-58897d9998-jqhqr\" (UID: \"0f3fd4f4-5ed4-40de-8632-65184569538c\") " pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.055532 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5f59867a-c7d1-42b8-8648-d50a179e80d0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.072468 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwnhs\" (UniqueName: \"kubernetes.io/projected/c4c40a07-5939-4d83-83af-e22f773ee511-kube-api-access-lwnhs\") pod \"oauth-openshift-558db77b4-nd2j6\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.080665 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.080837 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.58081007 +0000 UTC m=+153.775646119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.081189 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.081670 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.581657953 +0000 UTC m=+153.776494102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.088814 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs5mp\" (UniqueName: \"kubernetes.io/projected/57cd71b3-f7c5-49c5-9588-cd3107e1371d-kube-api-access-fs5mp\") pod \"kube-storage-version-migrator-operator-b67b599dd-8cw25\" (UID: \"57cd71b3-f7c5-49c5-9588-cd3107e1371d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.106148 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzv29\" (UniqueName: \"kubernetes.io/projected/c7e8cd6f-3f04-4ddd-9f7e-ff839255093d-kube-api-access-jzv29\") pod \"authentication-operator-69f744f599-bxn2h\" (UID: \"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.125907 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nld4x\" (UniqueName: \"kubernetes.io/projected/98d80369-7e37-4be7-beea-22dc5ef528c3-kube-api-access-nld4x\") pod \"controller-manager-879f6c89f-9bxjm\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.130583 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.149413 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.156088 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4sxx\" (UniqueName: \"kubernetes.io/projected/2678c8d7-202c-48b9-a5ec-cb56277af3a6-kube-api-access-j4sxx\") pod \"openshift-config-operator-7777fb866f-fb7dm\" (UID: \"2678c8d7-202c-48b9-a5ec-cb56277af3a6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.168734 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xkkr\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-kube-api-access-9xkkr\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.173735 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.181342 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.182462 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.182686 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.682653302 +0000 UTC m=+153.877489371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.182961 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.183635 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.683615258 +0000 UTC m=+153.878451327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.190898 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brf8t\" (UniqueName: \"kubernetes.io/projected/1e42c621-3be4-4655-afc9-2a55a4e548b1-kube-api-access-brf8t\") pod \"machine-approver-56656f9798-dvpb2\" (UID: \"1e42c621-3be4-4655-afc9-2a55a4e548b1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.200398 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.208939 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zkmv\" (UniqueName: \"kubernetes.io/projected/ddfac2cb-f987-4870-9453-afd3dce44113-kube-api-access-4zkmv\") pod \"apiserver-7bbb656c7d-crzrs\" (UID: \"ddfac2cb-f987-4870-9453-afd3dce44113\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.220383 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.234424 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-bound-sa-token\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.235495 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.262714 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.263410 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbb82\" (UniqueName: \"kubernetes.io/projected/5f59867a-c7d1-42b8-8648-d50a179e80d0-kube-api-access-rbb82\") pod \"cluster-image-registry-operator-dc59b4c8b-p2gm7\" (UID: \"5f59867a-c7d1-42b8-8648-d50a179e80d0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.280867 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22sm6\" (UniqueName: \"kubernetes.io/projected/40b8f502-e114-4817-82dd-fc33d6085119-kube-api-access-22sm6\") pod \"router-default-5444994796-gtvxw\" (UID: \"40b8f502-e114-4817-82dd-fc33d6085119\") " pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.284660 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.285060 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.78504813 +0000 UTC m=+153.979884179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.296074 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd94t\" (UniqueName: \"kubernetes.io/projected/a46886d7-03de-402d-9ea6-5b7c9d7f14c9-kube-api-access-wd94t\") pod \"machine-config-controller-84d6567774-hz2zl\" (UID: \"a46886d7-03de-402d-9ea6-5b7c9d7f14c9\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.296968 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.309871 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjbd6\" (UniqueName: \"kubernetes.io/projected/14a5a02a-4af8-4794-af5d-4a12c4098e3a-kube-api-access-vjbd6\") pod \"downloads-7954f5f757-wl4kz\" (UID: \"14a5a02a-4af8-4794-af5d-4a12c4098e3a\") " pod="openshift-console/downloads-7954f5f757-wl4kz" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.332813 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk9sx\" (UniqueName: \"kubernetes.io/projected/7c38baaf-c316-46a7-a224-9ae49128bc6f-kube-api-access-gk9sx\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.354970 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2skk\" (UniqueName: \"kubernetes.io/projected/02afee9d-7374-42ac-bef1-a2b321cb6418-kube-api-access-n2skk\") pod \"dns-operator-744455d44c-vbbs7\" (UID: \"02afee9d-7374-42ac-bef1-a2b321cb6418\") " pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.370631 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.376169 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4x9r\" (UniqueName: \"kubernetes.io/projected/37e0b2a3-df5f-4e4f-8146-ab0c30593448-kube-api-access-h4x9r\") pod \"machine-config-operator-74547568cd-dq59g\" (UID: \"37e0b2a3-df5f-4e4f-8146-ab0c30593448\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.387073 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.388026 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.888014082 +0000 UTC m=+154.082850131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.391912 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.396842 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5gqg\" (UniqueName: \"kubernetes.io/projected/8e9e25a5-cefa-401e-9126-562dc4a5c277-kube-api-access-w5gqg\") pod \"service-ca-9c57cc56f-v255n\" (UID: \"8e9e25a5-cefa-401e-9126-562dc4a5c277\") " pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.399747 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nd2j6"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.410871 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fef276de-096a-4b26-afdc-5a946f8a9502-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7lxx5\" (UID: \"fef276de-096a-4b26-afdc-5a946f8a9502\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.433587 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78vmh\" (UniqueName: \"kubernetes.io/projected/76b558c0-3be2-4d0d-af06-7cec72d518c3-kube-api-access-78vmh\") pod \"migrator-59844c95c7-8qz7f\" (UID: \"76b558c0-3be2-4d0d-af06-7cec72d518c3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.447302 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5acb54a3-8f1a-4d97-8200-1da490906e99-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-499gw\" (UID: \"5acb54a3-8f1a-4d97-8200-1da490906e99\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.454229 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" event={"ID":"b4de841e-a559-4470-afe3-8173d516d899","Type":"ContainerStarted","Data":"cf938a147554e26e1df5feb6aacf148bac8f095741037965b3a07fd6f2b54313"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.454270 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" event={"ID":"b4de841e-a559-4470-afe3-8173d516d899","Type":"ContainerStarted","Data":"ba1cd339fe42035e7bd277de57836d31e8f31d74fe591c90755b97b1de38cef4"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.455094 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.457328 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" event={"ID":"c4c40a07-5939-4d83-83af-e22f773ee511","Type":"ContainerStarted","Data":"d2a8eddfa669d16426be47cc1c342eb9b84d6a80fcca13fc3d3aca9f93b8f29b"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.467146 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd7jv\" (UniqueName: \"kubernetes.io/projected/a7b98640-cfde-41a4-81c1-a44d16db3219-kube-api-access-gd7jv\") pod \"csi-hostpathplugin-zs972\" (UID: \"a7b98640-cfde-41a4-81c1-a44d16db3219\") " pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.475256 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29525760-6k6m7" event={"ID":"bcd2fc23-8170-4bb9-90a2-5909cbea778a","Type":"ContainerStarted","Data":"2e87b0a66e13d66bc4cd81bad31c0901488a963878574e3d0cbeffee8e20dd04"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.479999 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gtvxw" event={"ID":"40b8f502-e114-4817-82dd-fc33d6085119","Type":"ContainerStarted","Data":"478b0eb48b43916cf4adba73616ebeb8634acd920fbceee2b280e0954f5e18b5"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.481364 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" event={"ID":"d2450fd9-6aae-4645-bcc7-94a67c7e0fef","Type":"ContainerStarted","Data":"5b9e06552e6be868155447b35b530fa3055dfa246a11ae7154d43646ac7fd577"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.481383 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" event={"ID":"d2450fd9-6aae-4645-bcc7-94a67c7e0fef","Type":"ContainerStarted","Data":"706f1aeaca434f0aa8f4bda64f6083c4cbb7e336258a35aed4e36a13ef54399a"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.481393 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" event={"ID":"d2450fd9-6aae-4645-bcc7-94a67c7e0fef","Type":"ContainerStarted","Data":"1ac669133129d132d2c670f00e8b6fc42177a6ea966327422dcde710febc1e0c"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.482876 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" event={"ID":"1e42c621-3be4-4655-afc9-2a55a4e548b1","Type":"ContainerStarted","Data":"585331d89fb38e439e3a6ab86aa139d923724bac3f9d3ae23d73b1409bc691ea"} Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.488377 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.488492 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.988462237 +0000 UTC m=+154.183298286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.488779 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.488830 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgjzf\" (UniqueName: \"kubernetes.io/projected/97fa91c6-34c8-454a-af96-84f2b73f513f-kube-api-access-bgjzf\") pod \"service-ca-operator-777779d784-2fxqs\" (UID: \"97fa91c6-34c8-454a-af96-84f2b73f513f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.489081 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:09.989067293 +0000 UTC m=+154.183903342 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.507507 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.509476 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dml4g\" (UniqueName: \"kubernetes.io/projected/1b9fcc3c-6634-4a00-bb46-2f467c25da5e-kube-api-access-dml4g\") pod \"dns-default-8496w\" (UID: \"1b9fcc3c-6634-4a00-bb46-2f467c25da5e\") " pod="openshift-dns/dns-default-8496w" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.514635 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-wl4kz" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.537422 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tln59\" (UniqueName: \"kubernetes.io/projected/eb30ed35-d423-4c83-a8e3-3294db1fd87d-kube-api-access-tln59\") pod \"olm-operator-6b444d44fb-txxk5\" (UID: \"eb30ed35-d423-4c83-a8e3-3294db1fd87d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.547513 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.554574 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmbdx\" (UniqueName: \"kubernetes.io/projected/1c602074-9ed8-4385-a338-40fefebb4924-kube-api-access-xmbdx\") pod \"control-plane-machine-set-operator-78cbb6b69f-z57xn\" (UID: \"1c602074-9ed8-4385-a338-40fefebb4924\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.569040 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lrtn\" (UniqueName: \"kubernetes.io/projected/1e2d40b9-179b-473d-b639-6c2ba54f2815-kube-api-access-6lrtn\") pod \"route-controller-manager-6576b87f9c-qbhj9\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.583467 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.585109 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.589510 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jqhqr"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.595062 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.595264 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.095225844 +0000 UTC m=+154.290061893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.595721 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.597559 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.097542478 +0000 UTC m=+154.292378617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.609279 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp6xr\" (UniqueName: \"kubernetes.io/projected/4d498c49-4df5-4e03-839f-739dafa7086f-kube-api-access-fp6xr\") pod \"multus-admission-controller-857f4d67dd-7zf67\" (UID: \"4d498c49-4df5-4e03-839f-739dafa7086f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.610952 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.612627 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l49cw\" (UniqueName: \"kubernetes.io/projected/9bdcfe02-eae7-4fb6-be1d-ed4216affdea-kube-api-access-l49cw\") pod \"etcd-operator-b45778765-ddrd9\" (UID: \"9bdcfe02-eae7-4fb6-be1d-ed4216affdea\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.618716 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.626519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6hfr\" (UniqueName: \"kubernetes.io/projected/c84179f9-cfa5-4d80-a5ba-67831c067e0d-kube-api-access-s6hfr\") pod \"packageserver-d55dfcdfc-rmtsj\" (UID: \"c84179f9-cfa5-4d80-a5ba-67831c067e0d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.634360 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9bxjm"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.635483 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v255n" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.643851 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.648239 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.652471 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.654994 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hjcc\" (UniqueName: \"kubernetes.io/projected/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-kube-api-access-4hjcc\") pod \"collect-profiles-29525760-jsh4m\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.669415 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkb7v\" (UniqueName: \"kubernetes.io/projected/0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d-kube-api-access-xkb7v\") pod \"ingress-canary-6f5qm\" (UID: \"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d\") " pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.676197 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.684010 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.696531 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.696690 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.196666525 +0000 UTC m=+154.391502574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.696761 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.697193 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.197186561 +0000 UTC m=+154.392022600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.701415 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgjnd\" (UniqueName: \"kubernetes.io/projected/55531783-59c5-4d74-b4af-d9e89c0a8421-kube-api-access-tgjnd\") pod \"marketplace-operator-79b997595-gksdw\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.701680 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.707268 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkc4f\" (UniqueName: \"kubernetes.io/projected/1c36fa76-92b7-43f8-9e82-97d2506cfb20-kube-api-access-gkc4f\") pod \"package-server-manager-789f6589d5-sz9qr\" (UID: \"1c36fa76-92b7-43f8-9e82-97d2506cfb20\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.716899 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6f5qm" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.732912 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn5bm\" (UniqueName: \"kubernetes.io/projected/9e5606c6-cb98-484a-ac5a-746ef795372c-kube-api-access-kn5bm\") pod \"machine-config-server-4mg6w\" (UID: \"9e5606c6-cb98-484a-ac5a-746ef795372c\") " pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.735287 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8496w" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.751828 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce0240f0-b334-484f-8037-040359dda7f7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lgx7\" (UID: \"ce0240f0-b334-484f-8037-040359dda7f7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.755564 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-zs972" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.757420 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.770603 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7c38baaf-c316-46a7-a224-9ae49128bc6f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vfxk6\" (UID: \"7c38baaf-c316-46a7-a224-9ae49128bc6f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.785831 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxn2h"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.787333 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.792877 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkmbv\" (UniqueName: \"kubernetes.io/projected/3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec-kube-api-access-mkmbv\") pod \"catalog-operator-68c6474976-qt6kv\" (UID: \"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.797688 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.797820 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.297797449 +0000 UTC m=+154.492633498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.798390 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.799753 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.299738132 +0000 UTC m=+154.494574181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.821366 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.834854 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.839894 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.864137 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.878092 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.899774 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.904851 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.905083 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:09 crc kubenswrapper[4750]: E0220 00:11:09.905594 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.405575264 +0000 UTC m=+154.600411313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.932392 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-wl4kz"] Feb 20 00:11:09 crc kubenswrapper[4750]: I0220 00:11:09.961493 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.006845 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.007435 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.507423717 +0000 UTC m=+154.702259766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.010301 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.025434 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4mg6w" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.103977 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7"] Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.107922 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.108192 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.608168419 +0000 UTC m=+154.803004468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.108248 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.109012 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.608998973 +0000 UTC m=+154.803835022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.208771 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.208943 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.708895601 +0000 UTC m=+154.903731650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.209457 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.209984 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.709971521 +0000 UTC m=+154.904807570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.235850 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v255n"] Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.315018 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.316609 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.816588445 +0000 UTC m=+155.011424484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.321207 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl"] Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.422567 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.423077 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:10.923064215 +0000 UTC m=+155.117900264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.524208 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.524405 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.024380332 +0000 UTC m=+155.219216381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.534809 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5"] Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.622500 4750 patch_prober.go:28] interesting pod/console-operator-58897d9998-jqhqr container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.623037 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" podUID="0f3fd4f4-5ed4-40de-8632-65184569538c" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.625222 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.625754 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.125738461 +0000 UTC m=+155.320574510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.637964 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.638003 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.638065 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" event={"ID":"0f3fd4f4-5ed4-40de-8632-65184569538c","Type":"ContainerStarted","Data":"2ffbb612522a3aadd634d3f25959b6a2e1b040371c87a546d4ae776a36fab953"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.638086 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" event={"ID":"0f3fd4f4-5ed4-40de-8632-65184569538c","Type":"ContainerStarted","Data":"08fcea07107f65f0e2a6cbb025afa3c09b3b1aba053782c13fd4ab305239d9f3"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.638098 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" event={"ID":"c4c40a07-5939-4d83-83af-e22f773ee511","Type":"ContainerStarted","Data":"68f0334058ba4fd26024c15c83c64f230975ff080336f93bb46820f3921ea075"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.694614 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" event={"ID":"37e0b2a3-df5f-4e4f-8146-ab0c30593448","Type":"ContainerStarted","Data":"0caa9f8918a9facd245d08945158fcaed6940c2d59f11018f09c55995b928608"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.721563 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" event={"ID":"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d","Type":"ContainerStarted","Data":"eea5c735cabb7ee959974cef1c05d3f8e275f7b8610b63a284cbe54f2d2bc6ef"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.726323 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gtvxw" event={"ID":"40b8f502-e114-4817-82dd-fc33d6085119","Type":"ContainerStarted","Data":"90553f223bf38f7a26b1c1305b6fdda8c7da2a51e40b6d8d3e5605500a54b2a1"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.729301 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.731169 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.231151532 +0000 UTC m=+155.425987581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.735396 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" event={"ID":"1e42c621-3be4-4655-afc9-2a55a4e548b1","Type":"ContainerStarted","Data":"cbf14bf319d1223ae93d80527e36fd8230715c923e5846de629057f181c8b52e"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.742927 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" event={"ID":"a46886d7-03de-402d-9ea6-5b7c9d7f14c9","Type":"ContainerStarted","Data":"a3e0b2af4772d51ba292901a331211c475027ff70358f967e385d61e4cf5ab14"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.743996 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" event={"ID":"ddfac2cb-f987-4870-9453-afd3dce44113","Type":"ContainerStarted","Data":"cf13f1b54e705e186bcf9e137861f94cd3122b0b05ace409b065891a3e31133f"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.744852 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v255n" event={"ID":"8e9e25a5-cefa-401e-9126-562dc4a5c277","Type":"ContainerStarted","Data":"69c8d64bc00f69fa559d8f7b4976f94b280fffc698fc642e9bf8d5f795727eee"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.746483 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" event={"ID":"8a972fe8-3fc2-42d8-b697-5addfe880cbe","Type":"ContainerStarted","Data":"997b946dccba30dffcab081dc2f5db86c8989f379570180b60ab03aee75356f9"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.746506 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" event={"ID":"8a972fe8-3fc2-42d8-b697-5addfe880cbe","Type":"ContainerStarted","Data":"108c9dbc571ed1f0ae355aa55ce934d1fef920e9ac49b89c4b63f60b7972473d"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.757084 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" event={"ID":"57cd71b3-f7c5-49c5-9588-cd3107e1371d","Type":"ContainerStarted","Data":"05dd2cde4e16310e468da0e9f539260012887cda263747d30e9ffffbaf56e11f"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.757153 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" event={"ID":"57cd71b3-f7c5-49c5-9588-cd3107e1371d","Type":"ContainerStarted","Data":"0ed6ba209f7303c9b0b1b75617a222105336035883dcf7a990bcb479346650d5"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.764430 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" event={"ID":"2678c8d7-202c-48b9-a5ec-cb56277af3a6","Type":"ContainerStarted","Data":"c10c5c97dde8e46bcff9f71ee31cef33a5925df4625c4566d434c8ac3395bfaf"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.764477 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" event={"ID":"2678c8d7-202c-48b9-a5ec-cb56277af3a6","Type":"ContainerStarted","Data":"71e092cb91370cb7eb4834e03651ba0dc08aa3a973a92683dced0500370c9399"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.771052 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4mg6w" event={"ID":"9e5606c6-cb98-484a-ac5a-746ef795372c","Type":"ContainerStarted","Data":"22256a4a5aea5bcc032c5019560d87ade1282dde4bf844233e4f1bbcfc3874f4"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.775311 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" event={"ID":"98d80369-7e37-4be7-beea-22dc5ef528c3","Type":"ContainerStarted","Data":"2ae02951e720a8a425fc90b7f8ed83063b5f6eea4963d6c89c7649e3c01d5773"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.777032 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.782320 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wl4kz" event={"ID":"14a5a02a-4af8-4794-af5d-4a12c4098e3a","Type":"ContainerStarted","Data":"ad031a3387133f2d873e485b81eac3e06807ea02d9e0870519f7ff86e84c5f82"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.786392 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" event={"ID":"5f59867a-c7d1-42b8-8648-d50a179e80d0","Type":"ContainerStarted","Data":"ae3df2ab3bf9235728453b8b6e215b53b81c298965886ecb7f1920c9d64a3440"} Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.786673 4750 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9bxjm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.786821 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.831012 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.835939 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.335923095 +0000 UTC m=+155.530759144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.923143 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-gk5dc" podStartSLOduration=128.923111634 podStartE2EDuration="2m8.923111634s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:10.92110845 +0000 UTC m=+155.115944499" watchObservedRunningTime="2026-02-20 00:11:10.923111634 +0000 UTC m=+155.117947683" Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.935354 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:10 crc kubenswrapper[4750]: E0220 00:11:10.936712 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.436606876 +0000 UTC m=+155.631442935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:10 crc kubenswrapper[4750]: I0220 00:11:10.966007 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29525760-6k6m7" podStartSLOduration=129.965970554 podStartE2EDuration="2m9.965970554s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:10.965101029 +0000 UTC m=+155.159937078" watchObservedRunningTime="2026-02-20 00:11:10.965970554 +0000 UTC m=+155.160806603" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.037350 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.038502 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.538488899 +0000 UTC m=+155.733324948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.068330 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kfsfb" podStartSLOduration=129.06830926 podStartE2EDuration="2m9.06830926s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.039406905 +0000 UTC m=+155.234242954" watchObservedRunningTime="2026-02-20 00:11:11.06830926 +0000 UTC m=+155.263145309" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.068452 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kw26r" podStartSLOduration=130.068446023 podStartE2EDuration="2m10.068446023s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.066885231 +0000 UTC m=+155.261721280" watchObservedRunningTime="2026-02-20 00:11:11.068446023 +0000 UTC m=+155.263282072" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.154688 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.154785 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.654763149 +0000 UTC m=+155.849599198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.154903 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.155212 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.655204501 +0000 UTC m=+155.850040550 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.198349 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.204648 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.206541 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.215094 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-zs972"] Feb 20 00:11:11 crc kubenswrapper[4750]: W0220 00:11:11.248098 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc84179f9_cfa5_4d80_a5ba_67831c067e0d.slice/crio-e9c0e687ae39c2585d21c56e93f357c8e3bd6026d8fa4cb4f02b1db3fdc0a4bc WatchSource:0}: Error finding container e9c0e687ae39c2585d21c56e93f357c8e3bd6026d8fa4cb4f02b1db3fdc0a4bc: Status 404 returned error can't find the container with id e9c0e687ae39c2585d21c56e93f357c8e3bd6026d8fa4cb4f02b1db3fdc0a4bc Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.250930 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.260306 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.260567 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.260973 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.76092405 +0000 UTC m=+155.955760099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.261125 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.267873 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.76783908 +0000 UTC m=+155.962675129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.286204 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.303341 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vbbs7"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.329656 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6f5qm"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.332039 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8496w"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.334127 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.344086 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8mkww" podStartSLOduration=130.344068118 podStartE2EDuration="2m10.344068118s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.343226575 +0000 UTC m=+155.538062614" watchObservedRunningTime="2026-02-20 00:11:11.344068118 +0000 UTC m=+155.538904167" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.364778 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.365178 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.865154268 +0000 UTC m=+156.059990317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.375801 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.394562 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:11 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:11 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:11 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.394606 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.423510 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.426382 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7zf67"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.431077 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.451462 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gksdw"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.463467 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.466812 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.467071 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:11.967059262 +0000 UTC m=+156.161895311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: W0220 00:11:11.482734 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3808ff6f_a9a5_4e4c_96e9_7beff1eed7ec.slice/crio-7694ca135b53559412b52aeed8f4c350f44937a9baf7ab2d965eb53dd87fa40d WatchSource:0}: Error finding container 7694ca135b53559412b52aeed8f4c350f44937a9baf7ab2d965eb53dd87fa40d: Status 404 returned error can't find the container with id 7694ca135b53559412b52aeed8f4c350f44937a9baf7ab2d965eb53dd87fa40d Feb 20 00:11:11 crc kubenswrapper[4750]: W0220 00:11:11.491302 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55531783_59c5_4d74_b4af_d9e89c0a8421.slice/crio-644278be123dee14b92de7c080ace0415aa47238664a2e3dfe770ea435462a61 WatchSource:0}: Error finding container 644278be123dee14b92de7c080ace0415aa47238664a2e3dfe770ea435462a61: Status 404 returned error can't find the container with id 644278be123dee14b92de7c080ace0415aa47238664a2e3dfe770ea435462a61 Feb 20 00:11:11 crc kubenswrapper[4750]: W0220 00:11:11.493012 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c36fa76_92b7_43f8_9e82_97d2506cfb20.slice/crio-6b1985060e15b4daee02334664d6b9a8d194aa53ae44ba708f6d9eb50fc7b0c6 WatchSource:0}: Error finding container 6b1985060e15b4daee02334664d6b9a8d194aa53ae44ba708f6d9eb50fc7b0c6: Status 404 returned error can't find the container with id 6b1985060e15b4daee02334664d6b9a8d194aa53ae44ba708f6d9eb50fc7b0c6 Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.544274 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.550659 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ddrd9"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.567078 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.567507 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.067487895 +0000 UTC m=+156.262323954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.570849 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m"] Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.593654 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7"] Feb 20 00:11:11 crc kubenswrapper[4750]: W0220 00:11:11.624326 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4b5ba72_9b9f_4a70_83ea_dd486b10f5a4.slice/crio-5535ffa471341c0b4c218c6229bf56502dc0fe27b49cc19498de6cffecba0331 WatchSource:0}: Error finding container 5535ffa471341c0b4c218c6229bf56502dc0fe27b49cc19498de6cffecba0331: Status 404 returned error can't find the container with id 5535ffa471341c0b4c218c6229bf56502dc0fe27b49cc19498de6cffecba0331 Feb 20 00:11:11 crc kubenswrapper[4750]: W0220 00:11:11.638422 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce0240f0_b334_484f_8037_040359dda7f7.slice/crio-a1e2146470041457f26e3bc4bf51824e329463b683b0eb13886e91aa38fa0dc4 WatchSource:0}: Error finding container a1e2146470041457f26e3bc4bf51824e329463b683b0eb13886e91aa38fa0dc4: Status 404 returned error can't find the container with id a1e2146470041457f26e3bc4bf51824e329463b683b0eb13886e91aa38fa0dc4 Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.673629 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.674077 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.174062678 +0000 UTC m=+156.368898727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.678833 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" podStartSLOduration=130.678815549 podStartE2EDuration="2m10.678815549s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.673911614 +0000 UTC m=+155.868747693" watchObservedRunningTime="2026-02-20 00:11:11.678815549 +0000 UTC m=+155.873651598" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.710771 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" podStartSLOduration=130.710748797 podStartE2EDuration="2m10.710748797s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.708957579 +0000 UTC m=+155.903793628" watchObservedRunningTime="2026-02-20 00:11:11.710748797 +0000 UTC m=+155.905584846" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.777245 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.777392 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.277369851 +0000 UTC m=+156.472205900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.777505 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.777807 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.277796352 +0000 UTC m=+156.472632401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.820705 4750 generic.go:334] "Generic (PLEG): container finished" podID="2678c8d7-202c-48b9-a5ec-cb56277af3a6" containerID="c10c5c97dde8e46bcff9f71ee31cef33a5925df4625c4566d434c8ac3395bfaf" exitCode=0 Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.821180 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" event={"ID":"2678c8d7-202c-48b9-a5ec-cb56277af3a6","Type":"ContainerDied","Data":"c10c5c97dde8e46bcff9f71ee31cef33a5925df4625c4566d434c8ac3395bfaf"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.821251 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" event={"ID":"2678c8d7-202c-48b9-a5ec-cb56277af3a6","Type":"ContainerStarted","Data":"34c020b72ea9757a2f01e26274458e65287be0453f45ba1f1413da3575c0641f"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.821270 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.826301 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" podStartSLOduration=129.826271567 podStartE2EDuration="2m9.826271567s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.801489445 +0000 UTC m=+155.996325494" watchObservedRunningTime="2026-02-20 00:11:11.826271567 +0000 UTC m=+156.021107616" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.826699 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mn2nx" podStartSLOduration=130.826693019 podStartE2EDuration="2m10.826693019s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.825158896 +0000 UTC m=+156.019994945" watchObservedRunningTime="2026-02-20 00:11:11.826693019 +0000 UTC m=+156.021529068" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.834725 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs972" event={"ID":"a7b98640-cfde-41a4-81c1-a44d16db3219","Type":"ContainerStarted","Data":"8e1e6d0f788b7632e31f51a57ccd75343e9f4c288e11829625a9d911f0a7cbc0"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.843237 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" event={"ID":"02afee9d-7374-42ac-bef1-a2b321cb6418","Type":"ContainerStarted","Data":"6a9ec5ee99bd876b10ee280d61e46d780bf99494db06394a8fb99f215f1ba798"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.878082 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.879042 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" podStartSLOduration=130.879015858 podStartE2EDuration="2m10.879015858s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.878156704 +0000 UTC m=+156.072992743" watchObservedRunningTime="2026-02-20 00:11:11.879015858 +0000 UTC m=+156.073851907" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.879093 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.3790746 +0000 UTC m=+156.573910649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.886748 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" event={"ID":"1e42c621-3be4-4655-afc9-2a55a4e548b1","Type":"ContainerStarted","Data":"102458bda16372dd94d6f1c433ea633131beef58bb5d9d28d235597b5ccbf6d5"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.890528 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" event={"ID":"ce0240f0-b334-484f-8037-040359dda7f7","Type":"ContainerStarted","Data":"a1e2146470041457f26e3bc4bf51824e329463b683b0eb13886e91aa38fa0dc4"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.891566 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" event={"ID":"55531783-59c5-4d74-b4af-d9e89c0a8421","Type":"ContainerStarted","Data":"644278be123dee14b92de7c080ace0415aa47238664a2e3dfe770ea435462a61"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.898636 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-wl4kz" event={"ID":"14a5a02a-4af8-4794-af5d-4a12c4098e3a","Type":"ContainerStarted","Data":"f14ba4ee92f93ec250bf25dbd9ceb49cccccede762aa40723c790a1658d9e7e9"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.900160 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-wl4kz" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.904092 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v255n" event={"ID":"8e9e25a5-cefa-401e-9126-562dc4a5c277","Type":"ContainerStarted","Data":"157e6857ee08cf83b1c9689e0e7423cd596a43647227d2c4eecb7c61891ff5bf"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.905321 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" event={"ID":"1c36fa76-92b7-43f8-9e82-97d2506cfb20","Type":"ContainerStarted","Data":"6b1985060e15b4daee02334664d6b9a8d194aa53ae44ba708f6d9eb50fc7b0c6"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.908582 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4mg6w" event={"ID":"9e5606c6-cb98-484a-ac5a-746ef795372c","Type":"ContainerStarted","Data":"803ac0e9bff6723bb4e69a61bbe64257fe99d28f89f407618f7449c5c2b5827d"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.912538 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-wl4kz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.912590 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wl4kz" podUID="14a5a02a-4af8-4794-af5d-4a12c4098e3a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.914532 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" podStartSLOduration=130.914520815 podStartE2EDuration="2m10.914520815s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.912937422 +0000 UTC m=+156.107773491" watchObservedRunningTime="2026-02-20 00:11:11.914520815 +0000 UTC m=+156.109356864" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.923197 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" event={"ID":"1c602074-9ed8-4385-a338-40fefebb4924","Type":"ContainerStarted","Data":"c7f204f55266d5b21a8f66d5e9bcfa13e94e7fcb710ff8d136f6a3ceea52d95b"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.923235 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" event={"ID":"1c602074-9ed8-4385-a338-40fefebb4924","Type":"ContainerStarted","Data":"965172bcbd0ddc88f7a7f05989e4f0432ead826392d4ccecd6f086d3ec194cb2"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.932146 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" event={"ID":"1e2d40b9-179b-473d-b639-6c2ba54f2815","Type":"ContainerStarted","Data":"2198033d69e89e4e2057e3136a80c737bcab5dd97253ca6b0b4cc3955cd80ec7"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.949660 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" event={"ID":"97fa91c6-34c8-454a-af96-84f2b73f513f","Type":"ContainerStarted","Data":"ba86f23fdaf71a74ea05febb1997665f1294f53f42d33bf87ef12eb25d1a37ec"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.949698 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" event={"ID":"97fa91c6-34c8-454a-af96-84f2b73f513f","Type":"ContainerStarted","Data":"ce7b20385fc79b6c38adb80f38db939b635792296729ecb717c36793b6cd46a1"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.954445 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8cw25" podStartSLOduration=129.954426273 podStartE2EDuration="2m9.954426273s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:11.95285848 +0000 UTC m=+156.147694539" watchObservedRunningTime="2026-02-20 00:11:11.954426273 +0000 UTC m=+156.149262322" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.963810 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" event={"ID":"eb30ed35-d423-4c83-a8e3-3294db1fd87d","Type":"ContainerStarted","Data":"c7209a38c216cad716b70580f3afc988fce75d83267c3a6bebf1f916f77b5c31"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.963859 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" event={"ID":"eb30ed35-d423-4c83-a8e3-3294db1fd87d","Type":"ContainerStarted","Data":"0bee697b509a3f65fe06681da087e0ca019b7580992c4344eed870d207a20d57"} Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.964880 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.981731 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.981855 4750 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-txxk5 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.981908 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" podUID="eb30ed35-d423-4c83-a8e3-3294db1fd87d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 20 00:11:11 crc kubenswrapper[4750]: E0220 00:11:11.991656 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.491623317 +0000 UTC m=+156.686459366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:11 crc kubenswrapper[4750]: I0220 00:11:11.998015 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" event={"ID":"9bdcfe02-eae7-4fb6-be1d-ed4216affdea","Type":"ContainerStarted","Data":"d2dbfbf0d471ec61bb3e0112d10f3db6c479f3a24f799ed90338e9cbff76b76f"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.008728 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-gtvxw" podStartSLOduration=130.008700247 podStartE2EDuration="2m10.008700247s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.005672784 +0000 UTC m=+156.200508833" watchObservedRunningTime="2026-02-20 00:11:12.008700247 +0000 UTC m=+156.203536296" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.040309 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" event={"ID":"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec","Type":"ContainerStarted","Data":"7694ca135b53559412b52aeed8f4c350f44937a9baf7ab2d965eb53dd87fa40d"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.059545 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" podStartSLOduration=131.059525935 podStartE2EDuration="2m11.059525935s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.058222539 +0000 UTC m=+156.253058608" watchObservedRunningTime="2026-02-20 00:11:12.059525935 +0000 UTC m=+156.254361994" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.061772 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" event={"ID":"a46886d7-03de-402d-9ea6-5b7c9d7f14c9","Type":"ContainerStarted","Data":"cd3ec18573a60aa07e768e9ca5907664063650cb93e714b3cba381d43e40b1ab"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.061850 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" event={"ID":"a46886d7-03de-402d-9ea6-5b7c9d7f14c9","Type":"ContainerStarted","Data":"15382a5ecb833eb9836dd478f1213c7cc57f5a2b306278c632c98c7d2262588a"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.072569 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" event={"ID":"5acb54a3-8f1a-4d97-8200-1da490906e99","Type":"ContainerStarted","Data":"961cc2942729706f157ffec629824667e6ef2ea61b7ab2e3cde9f0699f2eaf12"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.109877 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z57xn" podStartSLOduration=130.109864461 podStartE2EDuration="2m10.109864461s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.108717348 +0000 UTC m=+156.303553397" watchObservedRunningTime="2026-02-20 00:11:12.109864461 +0000 UTC m=+156.304700510" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.124637 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.137103 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" event={"ID":"76b558c0-3be2-4d0d-af06-7cec72d518c3","Type":"ContainerStarted","Data":"35d46abe91007e3de841e0c7d07c47b48d49310eef0bfc75398b0ef108ab69fe"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.137161 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" event={"ID":"76b558c0-3be2-4d0d-af06-7cec72d518c3","Type":"ContainerStarted","Data":"a88cbd8a68b1034b1293e3508c393981ebfac942e9fa88a1dced387b9571ea0c"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.137343 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.137620 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.137678 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.637660906 +0000 UTC m=+156.832496955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.194278 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4mg6w" podStartSLOduration=6.194248383 podStartE2EDuration="6.194248383s" podCreationTimestamp="2026-02-20 00:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.192661038 +0000 UTC m=+156.387497097" watchObservedRunningTime="2026-02-20 00:11:12.194248383 +0000 UTC m=+156.389084432" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.199507 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2fxqs" podStartSLOduration=130.199486677 podStartE2EDuration="2m10.199486677s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.157058649 +0000 UTC m=+156.351894698" watchObservedRunningTime="2026-02-20 00:11:12.199486677 +0000 UTC m=+156.394322726" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.200953 4750 patch_prober.go:28] interesting pod/apiserver-76f77b778f-jlgbs container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]log ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]etcd ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/generic-apiserver-start-informers ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/max-in-flight-filter ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 20 00:11:12 crc kubenswrapper[4750]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 20 00:11:12 crc kubenswrapper[4750]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/project.openshift.io-projectcache ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/openshift.io-startinformers ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 20 00:11:12 crc kubenswrapper[4750]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 20 00:11:12 crc kubenswrapper[4750]: livez check failed Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.201106 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxn2h" event={"ID":"c7e8cd6f-3f04-4ddd-9f7e-ff839255093d","Type":"ContainerStarted","Data":"57d80d38b840f7f36e9897bbd1940f6c2f427ad3c3f8989cb612bd3b5a688bbc"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.201673 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" podUID="b4de841e-a559-4470-afe3-8173d516d899" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.215721 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" event={"ID":"7c38baaf-c316-46a7-a224-9ae49128bc6f","Type":"ContainerStarted","Data":"b2f9c8d99f7fc5a908f6a3f0bc97aa757bf029a48575d5aae5fda7b9733f26b4"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.230155 4750 generic.go:334] "Generic (PLEG): container finished" podID="ddfac2cb-f987-4870-9453-afd3dce44113" containerID="7a53d1cf0326f183344b33b4e8a7aeaf1280206ea816f954ba4bf53c16752619" exitCode=0 Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.230242 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" event={"ID":"ddfac2cb-f987-4870-9453-afd3dce44113","Type":"ContainerDied","Data":"7a53d1cf0326f183344b33b4e8a7aeaf1280206ea816f954ba4bf53c16752619"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.231283 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" podStartSLOduration=131.231270561 podStartE2EDuration="2m11.231270561s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.226289924 +0000 UTC m=+156.421125973" watchObservedRunningTime="2026-02-20 00:11:12.231270561 +0000 UTC m=+156.426106610" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.235879 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" event={"ID":"4d498c49-4df5-4e03-839f-739dafa7086f","Type":"ContainerStarted","Data":"de1fe73eabb67bb05ec409d3a39161b3d52263376b8723bd560569b6f53f7c84"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.249688 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.252799 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.752784703 +0000 UTC m=+156.947620752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.255447 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" event={"ID":"fef276de-096a-4b26-afdc-5a946f8a9502","Type":"ContainerStarted","Data":"6b68f3ce82cbf33217b55f704af1f860c1d2b0baa20c6bf1e7040ed30c62088d"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.266185 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6f5qm" event={"ID":"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d","Type":"ContainerStarted","Data":"3615a86a786ff476336cc6c24003ea1156b8282a5dae78ca7b104bbd1a568a41"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.266238 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6f5qm" event={"ID":"0b3d5ebd-a3f9-4070-ad31-147e48ba7e0d","Type":"ContainerStarted","Data":"11c7d2598fc834371f26f16755e3c01230a354bb8cf40635e810a520e9c103ec"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.273719 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-wl4kz" podStartSLOduration=131.273691298 podStartE2EDuration="2m11.273691298s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.248498045 +0000 UTC m=+156.443334104" watchObservedRunningTime="2026-02-20 00:11:12.273691298 +0000 UTC m=+156.468527347" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.273849 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" podStartSLOduration=130.273845272 podStartE2EDuration="2m10.273845272s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.271519709 +0000 UTC m=+156.466355758" watchObservedRunningTime="2026-02-20 00:11:12.273845272 +0000 UTC m=+156.468681321" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.330341 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-v255n" podStartSLOduration=130.330312547 podStartE2EDuration="2m10.330312547s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.31915787 +0000 UTC m=+156.513993919" watchObservedRunningTime="2026-02-20 00:11:12.330312547 +0000 UTC m=+156.525148586" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.338425 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" event={"ID":"37e0b2a3-df5f-4e4f-8146-ab0c30593448","Type":"ContainerStarted","Data":"63ad458a9f1c5f3e52a4ada06a6247e1b3ccde7d1a45ec65e85929aafd1af03d"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.338482 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" event={"ID":"37e0b2a3-df5f-4e4f-8146-ab0c30593448","Type":"ContainerStarted","Data":"75dfa9082befd21a8ed1fdee5b66ae2805c5b43a69ee8a33a7b84f5674911364"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.350727 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.351540 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.85152192 +0000 UTC m=+157.046357969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.370559 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-dvpb2" podStartSLOduration=131.370542324 podStartE2EDuration="2m11.370542324s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.368738814 +0000 UTC m=+156.563574873" watchObservedRunningTime="2026-02-20 00:11:12.370542324 +0000 UTC m=+156.565378373" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.384410 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:12 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:12 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:12 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.384867 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.403471 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" event={"ID":"c84179f9-cfa5-4d80-a5ba-67831c067e0d","Type":"ContainerStarted","Data":"76eab9d9ba98b1b011f3a2c7da5a2bdc40fa09e8c77651e69300f00373f841ae"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.403536 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" event={"ID":"c84179f9-cfa5-4d80-a5ba-67831c067e0d","Type":"ContainerStarted","Data":"e9c0e687ae39c2585d21c56e93f357c8e3bd6026d8fa4cb4f02b1db3fdc0a4bc"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.404736 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.406665 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-6f5qm" podStartSLOduration=6.406649197 podStartE2EDuration="6.406649197s" podCreationTimestamp="2026-02-20 00:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.404516608 +0000 UTC m=+156.599352657" watchObservedRunningTime="2026-02-20 00:11:12.406649197 +0000 UTC m=+156.601485246" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.422302 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8496w" event={"ID":"1b9fcc3c-6634-4a00-bb46-2f467c25da5e","Type":"ContainerStarted","Data":"a4d2ebc8f1006e6af46c01b0e82d47248b82bc681f20e126adf2449e611fe737"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.427912 4750 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rmtsj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.44:5443/healthz\": dial tcp 10.217.0.44:5443: connect: connection refused" start-of-body= Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.427972 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" podUID="c84179f9-cfa5-4d80-a5ba-67831c067e0d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.44:5443/healthz\": dial tcp 10.217.0.44:5443: connect: connection refused" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.428105 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" event={"ID":"98d80369-7e37-4be7-beea-22dc5ef528c3","Type":"ContainerStarted","Data":"cc27c4ae5466e9f2596e970bde1cf3d1e16e40223aab82c1679451087e27783f"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.428598 4750 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9bxjm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.428631 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.455054 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.458080 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:12.958064832 +0000 UTC m=+157.152900881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.473004 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p2gm7" event={"ID":"5f59867a-c7d1-42b8-8648-d50a179e80d0","Type":"ContainerStarted","Data":"f26e693a798dc0d2912888397a895db04fbbe844de9b6b61649f574a596e15ab"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.499892 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" event={"ID":"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4","Type":"ContainerStarted","Data":"5535ffa471341c0b4c218c6229bf56502dc0fe27b49cc19498de6cffecba0331"} Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.536930 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" podStartSLOduration=130.536916571 podStartE2EDuration="2m10.536916571s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.500239212 +0000 UTC m=+156.695075261" watchObservedRunningTime="2026-02-20 00:11:12.536916571 +0000 UTC m=+156.731752610" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.537018 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dq59g" podStartSLOduration=130.537015324 podStartE2EDuration="2m10.537015324s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.535809801 +0000 UTC m=+156.730645850" watchObservedRunningTime="2026-02-20 00:11:12.537015324 +0000 UTC m=+156.731851373" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.545915 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-jqhqr" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.558535 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.558813 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.058798804 +0000 UTC m=+157.253634853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.564681 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" podStartSLOduration=130.564669005 podStartE2EDuration="2m10.564669005s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.564007387 +0000 UTC m=+156.758843436" watchObservedRunningTime="2026-02-20 00:11:12.564669005 +0000 UTC m=+156.759505054" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.603153 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-hz2zl" podStartSLOduration=130.603139664 podStartE2EDuration="2m10.603139664s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.602783174 +0000 UTC m=+156.797619223" watchObservedRunningTime="2026-02-20 00:11:12.603139664 +0000 UTC m=+156.797975703" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.660527 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.661511 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" podStartSLOduration=131.66149222 podStartE2EDuration="2m11.66149222s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.65968035 +0000 UTC m=+156.854516399" watchObservedRunningTime="2026-02-20 00:11:12.66149222 +0000 UTC m=+156.856328269" Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.664463 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.164449521 +0000 UTC m=+157.359285570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.764556 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.765077 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.26505787 +0000 UTC m=+157.459893919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.868800 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.869382 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.36936621 +0000 UTC m=+157.564202259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:12 crc kubenswrapper[4750]: I0220 00:11:12.973269 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:12 crc kubenswrapper[4750]: E0220 00:11:12.973818 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.473780693 +0000 UTC m=+157.668616752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.075909 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.077168 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.577144457 +0000 UTC m=+157.771980496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.180105 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.180438 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.680366318 +0000 UTC m=+157.875202377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.180540 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.181442 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.681429767 +0000 UTC m=+157.876265816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.281839 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.282007 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.781979644 +0000 UTC m=+157.976815693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.282254 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.282588 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.782580711 +0000 UTC m=+157.977416760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.305859 4750 csr.go:261] certificate signing request csr-tfpbf is approved, waiting to be issued Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.311336 4750 csr.go:257] certificate signing request csr-tfpbf is issued Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.375267 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:13 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:13 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:13 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.375718 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.383606 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.383777 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.883740654 +0000 UTC m=+158.078576703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.383905 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.384380 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.884371362 +0000 UTC m=+158.079207411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.485013 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.485190 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.985159325 +0000 UTC m=+158.179995384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.485256 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.485578 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:13.985564156 +0000 UTC m=+158.180400205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.515439 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" event={"ID":"4d498c49-4df5-4e03-839f-739dafa7086f","Type":"ContainerStarted","Data":"b8531f437dc7a98d059b9295d0bab4ac0539e0a47b1b94e671550f74e3fe483e"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.515486 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" event={"ID":"4d498c49-4df5-4e03-839f-739dafa7086f","Type":"ContainerStarted","Data":"a65994402519d92e49a4820531d80d337ad0a0241b2de5df48d00d54d7daf048"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.525202 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" event={"ID":"1e2d40b9-179b-473d-b639-6c2ba54f2815","Type":"ContainerStarted","Data":"c041a573e3f30b4a7a873e741804e813a85bd2b69553ccab12bc641701e5b9d0"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.525482 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.527969 4750 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-qbhj9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.528038 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" podUID="1e2d40b9-179b-473d-b639-6c2ba54f2815" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.528494 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" event={"ID":"9bdcfe02-eae7-4fb6-be1d-ed4216affdea","Type":"ContainerStarted","Data":"119af436574a4ef204919adb63f404cc4a2dd1daa960af5300620e1107f6d398"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.532678 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs972" event={"ID":"a7b98640-cfde-41a4-81c1-a44d16db3219","Type":"ContainerStarted","Data":"c1ddfef5197af22aae1fde1804cc35f425b00630ea9e3910ac4d414691bc0128"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.536183 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8496w" event={"ID":"1b9fcc3c-6634-4a00-bb46-2f467c25da5e","Type":"ContainerStarted","Data":"32e29f6666f2d75f9fdf85f75c95a771cecfdd408cd9b4389437d2c223a72fd0"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.536255 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-8496w" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.536273 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8496w" event={"ID":"1b9fcc3c-6634-4a00-bb46-2f467c25da5e","Type":"ContainerStarted","Data":"a6e488b2dea6918f46b1180c5af1cb02cd999af6da0f5351415fc5ef38cf39cc"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.537466 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7zf67" podStartSLOduration=131.537450783 podStartE2EDuration="2m11.537450783s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.533168306 +0000 UTC m=+157.728004355" watchObservedRunningTime="2026-02-20 00:11:13.537450783 +0000 UTC m=+157.732286832" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.537781 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" podStartSLOduration=131.537775683 podStartE2EDuration="2m11.537775683s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:12.806970043 +0000 UTC m=+157.001806092" watchObservedRunningTime="2026-02-20 00:11:13.537775683 +0000 UTC m=+157.732611742" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.539828 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" event={"ID":"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4","Type":"ContainerStarted","Data":"fa8a03d97b0252a3d464c84f24d335c694b43a2c82d8601c7c2bfff97d90cad0"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.543041 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7lxx5" event={"ID":"fef276de-096a-4b26-afdc-5a946f8a9502","Type":"ContainerStarted","Data":"2c8d5e80be79c442492dc5492031a77e2da07d39360f110761f4cdc601905ba1"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.545360 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" event={"ID":"ddfac2cb-f987-4870-9453-afd3dce44113","Type":"ContainerStarted","Data":"f135acea857472fb4bd8215923aa6034ec31cb98c73828d9fd19d89958e2ed66"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.549032 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-499gw" event={"ID":"5acb54a3-8f1a-4d97-8200-1da490906e99","Type":"ContainerStarted","Data":"e88cf07c9565e5b87f136333c23b88904ae10050301d3d8ccb39abb573f6d9d6"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.554370 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" event={"ID":"3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec","Type":"ContainerStarted","Data":"1d36aba51f1460f277895394b8812a2ff9b31c9f9e37e0c1c94b1aceb6711885"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.554591 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.556800 4750 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qt6kv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.556864 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" podUID="3808ff6f-a9a5-4e4c-96e9-7beff1eed7ec" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.557552 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" event={"ID":"1c36fa76-92b7-43f8-9e82-97d2506cfb20","Type":"ContainerStarted","Data":"395f02735bb0119d27883cb5fc40fd62406cb2e40d67d9ae0376ab4663125592"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.557607 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" event={"ID":"1c36fa76-92b7-43f8-9e82-97d2506cfb20","Type":"ContainerStarted","Data":"29b25b2602c03c930d439ee39c82f167e2a5e677e68c19706e9d63c854a8f39c"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.557671 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.563164 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" event={"ID":"7c38baaf-c316-46a7-a224-9ae49128bc6f","Type":"ContainerStarted","Data":"294d9c5a6b1978a72b38b9849e7e6819c715e20d2c8d09a3eaa5fc32d0729eb4"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.563238 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" event={"ID":"7c38baaf-c316-46a7-a224-9ae49128bc6f","Type":"ContainerStarted","Data":"0134a03831da8bda7b392308dd273f305d530e2bc7a4fd3be31fb454ad95cf00"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.567146 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" event={"ID":"ce0240f0-b334-484f-8037-040359dda7f7","Type":"ContainerStarted","Data":"0faa5d274e2b8e373214eb12a80f4c3b43c6dffcb45566aa25435ab4bafa8175"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.569400 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-ddrd9" podStartSLOduration=131.569385343 podStartE2EDuration="2m11.569385343s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.568317083 +0000 UTC m=+157.763153142" watchObservedRunningTime="2026-02-20 00:11:13.569385343 +0000 UTC m=+157.764221392" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.575290 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" event={"ID":"76b558c0-3be2-4d0d-af06-7cec72d518c3","Type":"ContainerStarted","Data":"f3b02f1df42ef72a8728175a41898bfdfbaac0b568b5363061e7424e9e168a31"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.585798 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.587500 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" event={"ID":"02afee9d-7374-42ac-bef1-a2b321cb6418","Type":"ContainerStarted","Data":"e1d31fcec2c2cf7429be07b50dd26391a4dca4696f8f3f2f62c657a77faf09cd"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.587559 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" event={"ID":"02afee9d-7374-42ac-bef1-a2b321cb6418","Type":"ContainerStarted","Data":"57c5b91199880a63c62b81aaf871bf91879c6a544c2d2cf1c4bef5630c16e613"} Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.591358 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.091339327 +0000 UTC m=+158.286175376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.605821 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" event={"ID":"55531783-59c5-4d74-b4af-d9e89c0a8421","Type":"ContainerStarted","Data":"6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a"} Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.605874 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.609587 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-wl4kz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.609635 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wl4kz" podUID="14a5a02a-4af8-4794-af5d-4a12c4098e3a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.624811 4750 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gksdw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.624866 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" podUID="55531783-59c5-4d74-b4af-d9e89c0a8421" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.643976 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.644221 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" podStartSLOduration=131.644209042 podStartE2EDuration="2m11.644209042s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.64196959 +0000 UTC m=+157.836805639" watchObservedRunningTime="2026-02-20 00:11:13.644209042 +0000 UTC m=+157.839045091" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.660470 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-txxk5" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.686629 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" podStartSLOduration=131.686603028 podStartE2EDuration="2m11.686603028s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.683273507 +0000 UTC m=+157.878109556" watchObservedRunningTime="2026-02-20 00:11:13.686603028 +0000 UTC m=+157.881439077" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.690260 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.696105 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.196074468 +0000 UTC m=+158.390910517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.717418 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" podStartSLOduration=131.717382485 podStartE2EDuration="2m11.717382485s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.709309413 +0000 UTC m=+157.904145462" watchObservedRunningTime="2026-02-20 00:11:13.717382485 +0000 UTC m=+157.912218534" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.746180 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lgx7" podStartSLOduration=131.746159367 podStartE2EDuration="2m11.746159367s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.745510989 +0000 UTC m=+157.940347048" watchObservedRunningTime="2026-02-20 00:11:13.746159367 +0000 UTC m=+157.940995416" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.801369 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.801781 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.301761567 +0000 UTC m=+158.496597616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.851432 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vfxk6" podStartSLOduration=131.851409983 podStartE2EDuration="2m11.851409983s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.825028817 +0000 UTC m=+158.019864866" watchObservedRunningTime="2026-02-20 00:11:13.851409983 +0000 UTC m=+158.046246032" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.894991 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" podStartSLOduration=131.894972932 podStartE2EDuration="2m11.894972932s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.893024158 +0000 UTC m=+158.087860207" watchObservedRunningTime="2026-02-20 00:11:13.894972932 +0000 UTC m=+158.089808981" Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.902932 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:13 crc kubenswrapper[4750]: E0220 00:11:13.903430 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.403413774 +0000 UTC m=+158.598249823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:13 crc kubenswrapper[4750]: I0220 00:11:13.986988 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-8496w" podStartSLOduration=7.986973573 podStartE2EDuration="7.986973573s" podCreationTimestamp="2026-02-20 00:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.948270548 +0000 UTC m=+158.143106597" watchObservedRunningTime="2026-02-20 00:11:13.986973573 +0000 UTC m=+158.181809622" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.006650 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.006968 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.506951753 +0000 UTC m=+158.701787802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.026403 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" podStartSLOduration=132.026388218 podStartE2EDuration="2m12.026388218s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:13.987416295 +0000 UTC m=+158.182252344" watchObservedRunningTime="2026-02-20 00:11:14.026388218 +0000 UTC m=+158.221224267" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.079673 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vbbs7" podStartSLOduration=132.079656544 podStartE2EDuration="2m12.079656544s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:14.035452937 +0000 UTC m=+158.230288986" watchObservedRunningTime="2026-02-20 00:11:14.079656544 +0000 UTC m=+158.274492593" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.109150 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.109473 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.609461564 +0000 UTC m=+158.804297613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.163482 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8qz7f" podStartSLOduration=132.16346675 podStartE2EDuration="2m12.16346675s" podCreationTimestamp="2026-02-20 00:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:14.082665406 +0000 UTC m=+158.277501455" watchObservedRunningTime="2026-02-20 00:11:14.16346675 +0000 UTC m=+158.358302799" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.211623 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.211963 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.711948794 +0000 UTC m=+158.906784843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.264673 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.264719 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.265768 4750 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-crzrs container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.19:8443/livez\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.265805 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" podUID="ddfac2cb-f987-4870-9453-afd3dce44113" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.19:8443/livez\": dial tcp 10.217.0.19:8443: connect: connection refused" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.312896 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-20 00:06:13 +0000 UTC, rotation deadline is 2026-11-23 21:39:45.728661969 +0000 UTC Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.312932 4750 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6645h28m31.415731877s for next certificate rotation Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.313474 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.313754 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.813745435 +0000 UTC m=+159.008581484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.380292 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:14 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:14 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:14 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.380349 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.414867 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.415030 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.915007242 +0000 UTC m=+159.109843291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.415140 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.415518 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:14.915502885 +0000 UTC m=+159.110338934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.516371 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.516515 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.016488234 +0000 UTC m=+159.211324273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.516633 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.516926 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.016912236 +0000 UTC m=+159.211748285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.608860 4750 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rmtsj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.44:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.609228 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" podUID="c84179f9-cfa5-4d80-a5ba-67831c067e0d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.44:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.609670 4750 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gksdw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.609712 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" podUID="55531783-59c5-4d74-b4af-d9e89c0a8421" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.617649 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.617824 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.117798712 +0000 UTC m=+159.312634761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.618662 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.619257 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.119233482 +0000 UTC m=+159.314069531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.623131 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.643912 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qt6kv" Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.719368 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.720645 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.220627031 +0000 UTC m=+159.415463080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.821560 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.821989 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.32197267 +0000 UTC m=+159.516808719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.922700 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.922833 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.422814525 +0000 UTC m=+159.617650574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:14 crc kubenswrapper[4750]: I0220 00:11:14.922921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:14 crc kubenswrapper[4750]: E0220 00:11:14.923188 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.423179345 +0000 UTC m=+159.618015394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.024288 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.024476 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.524450901 +0000 UTC m=+159.719286950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.024558 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.024811 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.524802721 +0000 UTC m=+159.719638770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.125757 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.125953 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.625927515 +0000 UTC m=+159.820763564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.126007 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.126477 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.626470969 +0000 UTC m=+159.821307018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.195812 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fb7dm" Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.227497 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.228182 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.728090575 +0000 UTC m=+159.922926624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.328869 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.329381 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.829363942 +0000 UTC m=+160.024199991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.380411 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:15 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:15 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:15 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.380470 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.430023 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.430207 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.930179426 +0000 UTC m=+160.125015475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.430341 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.431242 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:15.930651239 +0000 UTC m=+160.125487288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.534981 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.535488 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:16.035469934 +0000 UTC m=+160.230305983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.609386 4750 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rmtsj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.44:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.609447 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" podUID="c84179f9-cfa5-4d80-a5ba-67831c067e0d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.44:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.615691 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs972" event={"ID":"a7b98640-cfde-41a4-81c1-a44d16db3219","Type":"ContainerStarted","Data":"0a402161dafd9c73c08d7339dcd31087110cb8ee0700affd991ff8cf21da1fba"} Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.615783 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs972" event={"ID":"a7b98640-cfde-41a4-81c1-a44d16db3219","Type":"ContainerStarted","Data":"a641f872138c7be3dee8ec60f19b92e63bd5ae84df63db37eec42c3fae50b2a8"} Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.637001 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.637801 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:16.137784689 +0000 UTC m=+160.332620738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.708400 4750 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.738388 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.738498 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 00:11:16.23848138 +0000 UTC m=+160.433317429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.738771 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: E0220 00:11:15.739027 4750 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 00:11:16.239020244 +0000 UTC m=+160.433856293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5r77d" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.746190 4750 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-20T00:11:15.708425482Z","Handler":null,"Name":""} Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.761404 4750 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.761448 4750 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.840514 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.859552 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.942511 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.948517 4750 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.948549 4750 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:15 crc kubenswrapper[4750]: I0220 00:11:15.989346 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5r77d\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.017049 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-62qrs"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.018325 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.026023 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.034565 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62qrs"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.044536 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-catalog-content\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.044609 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vklkw\" (UniqueName: \"kubernetes.io/projected/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-kube-api-access-vklkw\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.044766 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-utilities\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.060219 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.147466 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-catalog-content\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.147551 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vklkw\" (UniqueName: \"kubernetes.io/projected/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-kube-api-access-vklkw\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.147609 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-utilities\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.147991 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-utilities\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.148494 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-catalog-content\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.177275 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vklkw\" (UniqueName: \"kubernetes.io/projected/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-kube-api-access-vklkw\") pod \"certified-operators-62qrs\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.212578 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rgp2g"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.213792 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.215977 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.227187 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgp2g"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.248610 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-catalog-content\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.248668 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-utilities\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.248712 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qqzn\" (UniqueName: \"kubernetes.io/projected/9ce86efc-8081-4418-ac52-06ce3348bc69-kube-api-access-5qqzn\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.323359 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5r77d"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.335909 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:11:16 crc kubenswrapper[4750]: W0220 00:11:16.337016 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d7c67d4_013f_494c_88dc_079a3accf05b.slice/crio-d3596ec8045423b04fc5289d0279b810bab12bc707ac4fbedc609ee91a94c1c5 WatchSource:0}: Error finding container d3596ec8045423b04fc5289d0279b810bab12bc707ac4fbedc609ee91a94c1c5: Status 404 returned error can't find the container with id d3596ec8045423b04fc5289d0279b810bab12bc707ac4fbedc609ee91a94c1c5 Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.349436 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-catalog-content\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.349493 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-utilities\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.349526 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qqzn\" (UniqueName: \"kubernetes.io/projected/9ce86efc-8081-4418-ac52-06ce3348bc69-kube-api-access-5qqzn\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.349885 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-catalog-content\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.349922 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-utilities\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.367462 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qqzn\" (UniqueName: \"kubernetes.io/projected/9ce86efc-8081-4418-ac52-06ce3348bc69-kube-api-access-5qqzn\") pod \"community-operators-rgp2g\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.374844 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:16 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:16 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:16 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.374906 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.408614 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gjlwq"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.409607 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.418554 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjlwq"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.450314 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xqh5\" (UniqueName: \"kubernetes.io/projected/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-kube-api-access-8xqh5\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.450381 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-catalog-content\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.450434 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-utilities\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.544702 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.552104 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-utilities\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.552571 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xqh5\" (UniqueName: \"kubernetes.io/projected/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-kube-api-access-8xqh5\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.552637 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-catalog-content\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.552755 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-utilities\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.553078 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-catalog-content\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.588891 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xqh5\" (UniqueName: \"kubernetes.io/projected/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-kube-api-access-8xqh5\") pod \"certified-operators-gjlwq\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.589633 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.631448 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5cq9b"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.643094 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5cq9b"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.643299 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.650190 4750 generic.go:334] "Generic (PLEG): container finished" podID="b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4" containerID="fa8a03d97b0252a3d464c84f24d335c694b43a2c82d8601c7c2bfff97d90cad0" exitCode=0 Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.650247 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" event={"ID":"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4","Type":"ContainerDied","Data":"fa8a03d97b0252a3d464c84f24d335c694b43a2c82d8601c7c2bfff97d90cad0"} Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.653520 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-utilities\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.653570 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-catalog-content\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.653702 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sjhf\" (UniqueName: \"kubernetes.io/projected/ac3d2b05-2e15-4058-a6f1-1ef965755b02-kube-api-access-2sjhf\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.657358 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-zs972" event={"ID":"a7b98640-cfde-41a4-81c1-a44d16db3219","Type":"ContainerStarted","Data":"f264b507e7c835027c164b43bf3d8cd4c687c45f6daa5abaf1626b02c6bca24d"} Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.668365 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" event={"ID":"5d7c67d4-013f-494c-88dc-079a3accf05b","Type":"ContainerStarted","Data":"d3596ec8045423b04fc5289d0279b810bab12bc707ac4fbedc609ee91a94c1c5"} Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.690280 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-zs972" podStartSLOduration=10.690262390000001 podStartE2EDuration="10.69026239s" podCreationTimestamp="2026-02-20 00:11:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:16.689740336 +0000 UTC m=+160.884576395" watchObservedRunningTime="2026-02-20 00:11:16.69026239 +0000 UTC m=+160.885098439" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.719867 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62qrs"] Feb 20 00:11:16 crc kubenswrapper[4750]: W0220 00:11:16.721627 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a63cf32_f756_45bb_ab8d_1e3aff3d2165.slice/crio-7579ca9463da1cd7d5bed2a0949beefc7ac468dadc690f774c9fcc7728e503c0 WatchSource:0}: Error finding container 7579ca9463da1cd7d5bed2a0949beefc7ac468dadc690f774c9fcc7728e503c0: Status 404 returned error can't find the container with id 7579ca9463da1cd7d5bed2a0949beefc7ac468dadc690f774c9fcc7728e503c0 Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.732593 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.754909 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sjhf\" (UniqueName: \"kubernetes.io/projected/ac3d2b05-2e15-4058-a6f1-1ef965755b02-kube-api-access-2sjhf\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.755072 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-utilities\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.755256 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-catalog-content\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.755909 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-catalog-content\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.757398 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-utilities\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.778225 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rgp2g"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.781151 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sjhf\" (UniqueName: \"kubernetes.io/projected/ac3d2b05-2e15-4058-a6f1-1ef965755b02-kube-api-access-2sjhf\") pod \"community-operators-5cq9b\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:16 crc kubenswrapper[4750]: W0220 00:11:16.782812 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ce86efc_8081_4418_ac52_06ce3348bc69.slice/crio-30cda0254c35d0de23abc101c09b14a4427574f52e0ba4ff3fd6bec3575eb845 WatchSource:0}: Error finding container 30cda0254c35d0de23abc101c09b14a4427574f52e0ba4ff3fd6bec3575eb845: Status 404 returned error can't find the container with id 30cda0254c35d0de23abc101c09b14a4427574f52e0ba4ff3fd6bec3575eb845 Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.965005 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjlwq"] Feb 20 00:11:16 crc kubenswrapper[4750]: I0220 00:11:16.966728 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.091158 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.094851 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-jlgbs" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.198780 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5cq9b"] Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.374071 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:17 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:17 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:17 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.374638 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.675907 4750 generic.go:334] "Generic (PLEG): container finished" podID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerID="2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317" exitCode=0 Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.676237 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62qrs" event={"ID":"3a63cf32-f756-45bb-ab8d-1e3aff3d2165","Type":"ContainerDied","Data":"2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.676280 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62qrs" event={"ID":"3a63cf32-f756-45bb-ab8d-1e3aff3d2165","Type":"ContainerStarted","Data":"7579ca9463da1cd7d5bed2a0949beefc7ac468dadc690f774c9fcc7728e503c0"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.677945 4750 generic.go:334] "Generic (PLEG): container finished" podID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerID="c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db" exitCode=0 Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.677995 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjlwq" event={"ID":"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7","Type":"ContainerDied","Data":"c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.678054 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjlwq" event={"ID":"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7","Type":"ContainerStarted","Data":"eb7daec71d0b615c7df4c4eeee250fbec344005cbc561c3ab947529d5163699b"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.678370 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.683743 4750 generic.go:334] "Generic (PLEG): container finished" podID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerID="b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5" exitCode=0 Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.684162 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5cq9b" event={"ID":"ac3d2b05-2e15-4058-a6f1-1ef965755b02","Type":"ContainerDied","Data":"b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.684206 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5cq9b" event={"ID":"ac3d2b05-2e15-4058-a6f1-1ef965755b02","Type":"ContainerStarted","Data":"de3feaccd4a83250e0b976aafd730794df8be48006f3badfa156163dd9b65c23"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.686985 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.687162 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.688542 4750 patch_prober.go:28] interesting pod/console-f9d7485db-kw26r container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.688589 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kw26r" podUID="3fc4787c-ec22-4961-be53-906fdd27c5f2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.689380 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" event={"ID":"5d7c67d4-013f-494c-88dc-079a3accf05b","Type":"ContainerStarted","Data":"7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.689519 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.693820 4750 generic.go:334] "Generic (PLEG): container finished" podID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerID="b43e1a9380ab0c28a54b7d6de786e862061af00435e5a5bec32c2d73d7efac4c" exitCode=0 Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.694790 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgp2g" event={"ID":"9ce86efc-8081-4418-ac52-06ce3348bc69","Type":"ContainerDied","Data":"b43e1a9380ab0c28a54b7d6de786e862061af00435e5a5bec32c2d73d7efac4c"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.694832 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgp2g" event={"ID":"9ce86efc-8081-4418-ac52-06ce3348bc69","Type":"ContainerStarted","Data":"30cda0254c35d0de23abc101c09b14a4427574f52e0ba4ff3fd6bec3575eb845"} Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.778751 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" podStartSLOduration=136.778731331 podStartE2EDuration="2m16.778731331s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:17.775169644 +0000 UTC m=+161.970005693" watchObservedRunningTime="2026-02-20 00:11:17.778731331 +0000 UTC m=+161.973567390" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.935310 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.971213 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hjcc\" (UniqueName: \"kubernetes.io/projected/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-kube-api-access-4hjcc\") pod \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.971309 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-secret-volume\") pod \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.971356 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-config-volume\") pod \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\" (UID: \"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4\") " Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.972025 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-config-volume" (OuterVolumeSpecName: "config-volume") pod "b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4" (UID: "b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.977678 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-kube-api-access-4hjcc" (OuterVolumeSpecName: "kube-api-access-4hjcc") pod "b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4" (UID: "b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4"). InnerVolumeSpecName "kube-api-access-4hjcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.978035 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4" (UID: "b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.987357 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 00:11:17 crc kubenswrapper[4750]: E0220 00:11:17.987587 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4" containerName="collect-profiles" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.987607 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4" containerName="collect-profiles" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.987716 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4" containerName="collect-profiles" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.988047 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.989729 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.989939 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 20 00:11:17 crc kubenswrapper[4750]: I0220 00:11:17.996385 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.072802 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9641156-dece-4bc9-8734-0b77e320f013-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.072942 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9641156-dece-4bc9-8734-0b77e320f013-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.073074 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.073090 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.073105 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hjcc\" (UniqueName: \"kubernetes.io/projected/b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4-kube-api-access-4hjcc\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.174337 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9641156-dece-4bc9-8734-0b77e320f013-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.174449 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9641156-dece-4bc9-8734-0b77e320f013-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.174540 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9641156-dece-4bc9-8734-0b77e320f013-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.192641 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9641156-dece-4bc9-8734-0b77e320f013-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.213545 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z2vd9"] Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.214509 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.216465 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.229008 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2vd9"] Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.275958 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zck6g\" (UniqueName: \"kubernetes.io/projected/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-kube-api-access-zck6g\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.276051 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-catalog-content\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.276135 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-utilities\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.302006 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.375052 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:18 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:18 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:18 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.375144 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.377721 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-utilities\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.377814 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zck6g\" (UniqueName: \"kubernetes.io/projected/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-kube-api-access-zck6g\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.377936 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-catalog-content\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.378991 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-catalog-content\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.379565 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-utilities\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.413272 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zck6g\" (UniqueName: \"kubernetes.io/projected/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-kube-api-access-zck6g\") pod \"redhat-marketplace-z2vd9\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.532585 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.610299 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbjg"] Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.611443 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.620509 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbjg"] Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.682582 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-catalog-content\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.682647 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjqm9\" (UniqueName: \"kubernetes.io/projected/9595c598-a6fe-45fc-a934-2a64ca8ef288-kube-api-access-kjqm9\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.682679 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-utilities\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.708693 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.709226 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29525760-jsh4m" event={"ID":"b4b5ba72-9b9f-4a70-83ea-dd486b10f5a4","Type":"ContainerDied","Data":"5535ffa471341c0b4c218c6229bf56502dc0fe27b49cc19498de6cffecba0331"} Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.709268 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5535ffa471341c0b4c218c6229bf56502dc0fe27b49cc19498de6cffecba0331" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.735553 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.782144 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2vd9"] Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.784529 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-utilities\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.785244 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-utilities\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.785488 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-catalog-content\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.785549 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjqm9\" (UniqueName: \"kubernetes.io/projected/9595c598-a6fe-45fc-a934-2a64ca8ef288-kube-api-access-kjqm9\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.787171 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-catalog-content\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: W0220 00:11:18.805738 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e3cf328_65f5_417b_8bb4_35ff194e8cc4.slice/crio-a29a1c79128721128c764e69281513c8874538d70f672b3cc47820150bfb3f07 WatchSource:0}: Error finding container a29a1c79128721128c764e69281513c8874538d70f672b3cc47820150bfb3f07: Status 404 returned error can't find the container with id a29a1c79128721128c764e69281513c8874538d70f672b3cc47820150bfb3f07 Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.822028 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjqm9\" (UniqueName: \"kubernetes.io/projected/9595c598-a6fe-45fc-a934-2a64ca8ef288-kube-api-access-kjqm9\") pod \"redhat-marketplace-4zbjg\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:18 crc kubenswrapper[4750]: I0220 00:11:18.935999 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.212067 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sll9r"] Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.216350 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.230849 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.234385 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sll9r"] Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.270226 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.275277 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-crzrs" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.295554 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-catalog-content\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.295591 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-utilities\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.295733 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlgqf\" (UniqueName: \"kubernetes.io/projected/295b160a-bc39-40be-822d-fc4438895480-kube-api-access-wlgqf\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.371685 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.377282 4750 patch_prober.go:28] interesting pod/router-default-5444994796-gtvxw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 00:11:19 crc kubenswrapper[4750]: [-]has-synced failed: reason withheld Feb 20 00:11:19 crc kubenswrapper[4750]: [+]process-running ok Feb 20 00:11:19 crc kubenswrapper[4750]: healthz check failed Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.377361 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gtvxw" podUID="40b8f502-e114-4817-82dd-fc33d6085119" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.400032 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlgqf\" (UniqueName: \"kubernetes.io/projected/295b160a-bc39-40be-822d-fc4438895480-kube-api-access-wlgqf\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.400156 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-catalog-content\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.400177 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-utilities\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.400591 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-utilities\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.401666 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-catalog-content\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.423513 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlgqf\" (UniqueName: \"kubernetes.io/projected/295b160a-bc39-40be-822d-fc4438895480-kube-api-access-wlgqf\") pod \"redhat-operators-sll9r\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.490815 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbjg"] Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.516176 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-wl4kz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.516212 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-wl4kz" podUID="14a5a02a-4af8-4794-af5d-4a12c4098e3a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.516263 4750 patch_prober.go:28] interesting pod/downloads-7954f5f757-wl4kz container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.516277 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-wl4kz" podUID="14a5a02a-4af8-4794-af5d-4a12c4098e3a" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.567566 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.617808 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8v8d"] Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.618926 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.620862 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8v8d"] Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.688869 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmtsj" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.707707 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-utilities\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.708230 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-catalog-content\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.708310 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w5wt\" (UniqueName: \"kubernetes.io/projected/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-kube-api-access-5w5wt\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.746632 4750 generic.go:334] "Generic (PLEG): container finished" podID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerID="9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271" exitCode=0 Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.746688 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2vd9" event={"ID":"0e3cf328-65f5-417b-8bb4-35ff194e8cc4","Type":"ContainerDied","Data":"9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271"} Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.746714 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2vd9" event={"ID":"0e3cf328-65f5-417b-8bb4-35ff194e8cc4","Type":"ContainerStarted","Data":"a29a1c79128721128c764e69281513c8874538d70f672b3cc47820150bfb3f07"} Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.772773 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a9641156-dece-4bc9-8734-0b77e320f013","Type":"ContainerStarted","Data":"58955b0c409f53058d462c7a694d1feb1016751e4a3bef4bd2822367c444339f"} Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.772807 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a9641156-dece-4bc9-8734-0b77e320f013","Type":"ContainerStarted","Data":"318d023893da9a6fabe6a05ae00cd44665dea3d6c686ac418708a229ba93ce80"} Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.789233 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbjg" event={"ID":"9595c598-a6fe-45fc-a934-2a64ca8ef288","Type":"ContainerStarted","Data":"b71a858dba8de9842ed96380aeab2a9eaecb2a4d928e7d44c1cecebc4a26e3f8"} Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.809880 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w5wt\" (UniqueName: \"kubernetes.io/projected/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-kube-api-access-5w5wt\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.809994 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-utilities\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.810062 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-catalog-content\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.812215 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-utilities\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.812886 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-catalog-content\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.837379 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w5wt\" (UniqueName: \"kubernetes.io/projected/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-kube-api-access-5w5wt\") pod \"redhat-operators-g8v8d\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.908669 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.924197 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.924170668 podStartE2EDuration="2.924170668s" podCreationTimestamp="2026-02-20 00:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:19.799693044 +0000 UTC m=+163.994529093" watchObservedRunningTime="2026-02-20 00:11:19.924170668 +0000 UTC m=+164.119006717" Feb 20 00:11:19 crc kubenswrapper[4750]: I0220 00:11:19.976086 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.059700 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sll9r"] Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.305487 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8v8d"] Feb 20 00:11:20 crc kubenswrapper[4750]: W0220 00:11:20.353382 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode21f8aa9_f543_4f68_b241_fde6f0e6bb49.slice/crio-2a389c259fadf83772f601d5a047fec956873874f61d4307ca3f487639f4d68f WatchSource:0}: Error finding container 2a389c259fadf83772f601d5a047fec956873874f61d4307ca3f487639f4d68f: Status 404 returned error can't find the container with id 2a389c259fadf83772f601d5a047fec956873874f61d4307ca3f487639f4d68f Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.374628 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.380392 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gtvxw" Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.798247 4750 generic.go:334] "Generic (PLEG): container finished" podID="a9641156-dece-4bc9-8734-0b77e320f013" containerID="58955b0c409f53058d462c7a694d1feb1016751e4a3bef4bd2822367c444339f" exitCode=0 Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.798518 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a9641156-dece-4bc9-8734-0b77e320f013","Type":"ContainerDied","Data":"58955b0c409f53058d462c7a694d1feb1016751e4a3bef4bd2822367c444339f"} Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.800797 4750 generic.go:334] "Generic (PLEG): container finished" podID="295b160a-bc39-40be-822d-fc4438895480" containerID="f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb" exitCode=0 Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.800866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sll9r" event={"ID":"295b160a-bc39-40be-822d-fc4438895480","Type":"ContainerDied","Data":"f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb"} Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.800889 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sll9r" event={"ID":"295b160a-bc39-40be-822d-fc4438895480","Type":"ContainerStarted","Data":"e2d36dad879b2e53b716779d32d770d3d42cc7123a33f12877e1850bc7f131cf"} Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.803934 4750 generic.go:334] "Generic (PLEG): container finished" podID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerID="f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f" exitCode=0 Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.804571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbjg" event={"ID":"9595c598-a6fe-45fc-a934-2a64ca8ef288","Type":"ContainerDied","Data":"f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f"} Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.815312 4750 generic.go:334] "Generic (PLEG): container finished" podID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerID="f3c4eee74f8e7c7e8d8b74e47b77b3a4c38749577ddf297e570811825a96e30f" exitCode=0 Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.815549 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8v8d" event={"ID":"e21f8aa9-f543-4f68-b241-fde6f0e6bb49","Type":"ContainerDied","Data":"f3c4eee74f8e7c7e8d8b74e47b77b3a4c38749577ddf297e570811825a96e30f"} Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.815588 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8v8d" event={"ID":"e21f8aa9-f543-4f68-b241-fde6f0e6bb49","Type":"ContainerStarted","Data":"2a389c259fadf83772f601d5a047fec956873874f61d4307ca3f487639f4d68f"} Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.838537 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.840991 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.852137 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.872024 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.873535 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.942390 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:20 crc kubenswrapper[4750]: I0220 00:11:20.942595 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:21 crc kubenswrapper[4750]: I0220 00:11:21.044151 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:21 crc kubenswrapper[4750]: I0220 00:11:21.044220 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:21 crc kubenswrapper[4750]: I0220 00:11:21.044287 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:21 crc kubenswrapper[4750]: I0220 00:11:21.062636 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:21 crc kubenswrapper[4750]: I0220 00:11:21.181042 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:21 crc kubenswrapper[4750]: I0220 00:11:21.456281 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 00:11:21 crc kubenswrapper[4750]: I0220 00:11:21.898707 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7","Type":"ContainerStarted","Data":"e3824a6004d8d2c4eeb315617a236fba8634551fd5000ada92ef6b153114848e"} Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.242054 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.265470 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9641156-dece-4bc9-8734-0b77e320f013-kubelet-dir\") pod \"a9641156-dece-4bc9-8734-0b77e320f013\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.265554 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9641156-dece-4bc9-8734-0b77e320f013-kube-api-access\") pod \"a9641156-dece-4bc9-8734-0b77e320f013\" (UID: \"a9641156-dece-4bc9-8734-0b77e320f013\") " Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.265888 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9641156-dece-4bc9-8734-0b77e320f013-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a9641156-dece-4bc9-8734-0b77e320f013" (UID: "a9641156-dece-4bc9-8734-0b77e320f013"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.271615 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9641156-dece-4bc9-8734-0b77e320f013-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a9641156-dece-4bc9-8734-0b77e320f013" (UID: "a9641156-dece-4bc9-8734-0b77e320f013"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.367303 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a9641156-dece-4bc9-8734-0b77e320f013-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.367343 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a9641156-dece-4bc9-8734-0b77e320f013-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.968815 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7","Type":"ContainerStarted","Data":"e18790cacfd9e84690d7e43c4a967ddc519b52c88f3569fdf24599cbd36b9814"} Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.987369 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.987479 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a9641156-dece-4bc9-8734-0b77e320f013","Type":"ContainerDied","Data":"318d023893da9a6fabe6a05ae00cd44665dea3d6c686ac418708a229ba93ce80"} Feb 20 00:11:22 crc kubenswrapper[4750]: I0220 00:11:22.987523 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="318d023893da9a6fabe6a05ae00cd44665dea3d6c686ac418708a229ba93ce80" Feb 20 00:11:23 crc kubenswrapper[4750]: I0220 00:11:23.002362 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.002344842 podStartE2EDuration="3.002344842s" podCreationTimestamp="2026-02-20 00:11:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:22.999678668 +0000 UTC m=+167.194514717" watchObservedRunningTime="2026-02-20 00:11:23.002344842 +0000 UTC m=+167.197180891" Feb 20 00:11:24 crc kubenswrapper[4750]: I0220 00:11:23.999788 4750 generic.go:334] "Generic (PLEG): container finished" podID="ce4f9b76-9136-4deb-b4da-7d5c6e5efad7" containerID="e18790cacfd9e84690d7e43c4a967ddc519b52c88f3569fdf24599cbd36b9814" exitCode=0 Feb 20 00:11:24 crc kubenswrapper[4750]: I0220 00:11:24.000085 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7","Type":"ContainerDied","Data":"e18790cacfd9e84690d7e43c4a967ddc519b52c88f3569fdf24599cbd36b9814"} Feb 20 00:11:24 crc kubenswrapper[4750]: I0220 00:11:24.737871 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-8496w" Feb 20 00:11:25 crc kubenswrapper[4750]: I0220 00:11:25.430788 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:11:25 crc kubenswrapper[4750]: I0220 00:11:25.438519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fac6ee29-99b8-42e7-844d-30e68cbecad3-metrics-certs\") pod \"network-metrics-daemon-tcgh4\" (UID: \"fac6ee29-99b8-42e7-844d-30e68cbecad3\") " pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:11:25 crc kubenswrapper[4750]: I0220 00:11:25.482225 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tcgh4" Feb 20 00:11:26 crc kubenswrapper[4750]: I0220 00:11:26.670055 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:11:26 crc kubenswrapper[4750]: I0220 00:11:26.670106 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:11:27 crc kubenswrapper[4750]: I0220 00:11:27.686922 4750 patch_prober.go:28] interesting pod/console-f9d7485db-kw26r container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Feb 20 00:11:27 crc kubenswrapper[4750]: I0220 00:11:27.687250 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kw26r" podUID="3fc4787c-ec22-4961-be53-906fdd27c5f2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.32:8443/health\": dial tcp 10.217.0.32:8443: connect: connection refused" Feb 20 00:11:29 crc kubenswrapper[4750]: I0220 00:11:29.519642 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-wl4kz" Feb 20 00:11:29 crc kubenswrapper[4750]: I0220 00:11:29.975470 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9bxjm"] Feb 20 00:11:29 crc kubenswrapper[4750]: I0220 00:11:29.975676 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" containerID="cri-o://cc27c4ae5466e9f2596e970bde1cf3d1e16e40223aab82c1679451087e27783f" gracePeriod=30 Feb 20 00:11:29 crc kubenswrapper[4750]: I0220 00:11:29.996351 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9"] Feb 20 00:11:29 crc kubenswrapper[4750]: I0220 00:11:29.996866 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" podUID="1e2d40b9-179b-473d-b639-6c2ba54f2815" containerName="route-controller-manager" containerID="cri-o://c041a573e3f30b4a7a873e741804e813a85bd2b69553ccab12bc641701e5b9d0" gracePeriod=30 Feb 20 00:11:31 crc kubenswrapper[4750]: I0220 00:11:31.082508 4750 generic.go:334] "Generic (PLEG): container finished" podID="1e2d40b9-179b-473d-b639-6c2ba54f2815" containerID="c041a573e3f30b4a7a873e741804e813a85bd2b69553ccab12bc641701e5b9d0" exitCode=0 Feb 20 00:11:31 crc kubenswrapper[4750]: I0220 00:11:31.082571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" event={"ID":"1e2d40b9-179b-473d-b639-6c2ba54f2815","Type":"ContainerDied","Data":"c041a573e3f30b4a7a873e741804e813a85bd2b69553ccab12bc641701e5b9d0"} Feb 20 00:11:31 crc kubenswrapper[4750]: I0220 00:11:31.084207 4750 generic.go:334] "Generic (PLEG): container finished" podID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerID="cc27c4ae5466e9f2596e970bde1cf3d1e16e40223aab82c1679451087e27783f" exitCode=0 Feb 20 00:11:31 crc kubenswrapper[4750]: I0220 00:11:31.084260 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" event={"ID":"98d80369-7e37-4be7-beea-22dc5ef528c3","Type":"ContainerDied","Data":"cc27c4ae5466e9f2596e970bde1cf3d1e16e40223aab82c1679451087e27783f"} Feb 20 00:11:34 crc kubenswrapper[4750]: I0220 00:11:34.597891 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 00:11:36 crc kubenswrapper[4750]: I0220 00:11:36.066943 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:11:37 crc kubenswrapper[4750]: I0220 00:11:37.693494 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:37 crc kubenswrapper[4750]: I0220 00:11:37.698942 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kw26r" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.060111 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.068098 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.127830 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.128680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9" event={"ID":"1e2d40b9-179b-473d-b639-6c2ba54f2815","Type":"ContainerDied","Data":"2198033d69e89e4e2057e3136a80c737bcab5dd97253ca6b0b4cc3955cd80ec7"} Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.130152 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.130443 4750 scope.go:117] "RemoveContainer" containerID="c041a573e3f30b4a7a873e741804e813a85bd2b69553ccab12bc641701e5b9d0" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.130361 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7","Type":"ContainerDied","Data":"e3824a6004d8d2c4eeb315617a236fba8634551fd5000ada92ef6b153114848e"} Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.130849 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3824a6004d8d2c4eeb315617a236fba8634551fd5000ada92ef6b153114848e" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204072 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e2d40b9-179b-473d-b639-6c2ba54f2815-serving-cert\") pod \"1e2d40b9-179b-473d-b639-6c2ba54f2815\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204190 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kubelet-dir\") pod \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204226 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kube-api-access\") pod \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\" (UID: \"ce4f9b76-9136-4deb-b4da-7d5c6e5efad7\") " Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204335 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-config\") pod \"1e2d40b9-179b-473d-b639-6c2ba54f2815\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204371 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ce4f9b76-9136-4deb-b4da-7d5c6e5efad7" (UID: "ce4f9b76-9136-4deb-b4da-7d5c6e5efad7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204403 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-client-ca\") pod \"1e2d40b9-179b-473d-b639-6c2ba54f2815\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204487 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lrtn\" (UniqueName: \"kubernetes.io/projected/1e2d40b9-179b-473d-b639-6c2ba54f2815-kube-api-access-6lrtn\") pod \"1e2d40b9-179b-473d-b639-6c2ba54f2815\" (UID: \"1e2d40b9-179b-473d-b639-6c2ba54f2815\") " Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204850 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-client-ca" (OuterVolumeSpecName: "client-ca") pod "1e2d40b9-179b-473d-b639-6c2ba54f2815" (UID: "1e2d40b9-179b-473d-b639-6c2ba54f2815"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.204950 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.205575 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-config" (OuterVolumeSpecName: "config") pod "1e2d40b9-179b-473d-b639-6c2ba54f2815" (UID: "1e2d40b9-179b-473d-b639-6c2ba54f2815"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.209434 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ce4f9b76-9136-4deb-b4da-7d5c6e5efad7" (UID: "ce4f9b76-9136-4deb-b4da-7d5c6e5efad7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.209760 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e2d40b9-179b-473d-b639-6c2ba54f2815-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1e2d40b9-179b-473d-b639-6c2ba54f2815" (UID: "1e2d40b9-179b-473d-b639-6c2ba54f2815"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.210179 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e2d40b9-179b-473d-b639-6c2ba54f2815-kube-api-access-6lrtn" (OuterVolumeSpecName: "kube-api-access-6lrtn") pod "1e2d40b9-179b-473d-b639-6c2ba54f2815" (UID: "1e2d40b9-179b-473d-b639-6c2ba54f2815"). InnerVolumeSpecName "kube-api-access-6lrtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.306299 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.306351 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e2d40b9-179b-473d-b639-6c2ba54f2815-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.306372 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lrtn\" (UniqueName: \"kubernetes.io/projected/1e2d40b9-179b-473d-b639-6c2ba54f2815-kube-api-access-6lrtn\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.306390 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e2d40b9-179b-473d-b639-6c2ba54f2815-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.306407 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ce4f9b76-9136-4deb-b4da-7d5c6e5efad7-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.459535 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9"] Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.466560 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qbhj9"] Feb 20 00:11:38 crc kubenswrapper[4750]: I0220 00:11:38.584571 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e2d40b9-179b-473d-b639-6c2ba54f2815" path="/var/lib/kubelet/pods/1e2d40b9-179b-473d-b639-6c2ba54f2815/volumes" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.236364 4750 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9bxjm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.236753 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.601510 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm"] Feb 20 00:11:40 crc kubenswrapper[4750]: E0220 00:11:40.601743 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4f9b76-9136-4deb-b4da-7d5c6e5efad7" containerName="pruner" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.601755 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4f9b76-9136-4deb-b4da-7d5c6e5efad7" containerName="pruner" Feb 20 00:11:40 crc kubenswrapper[4750]: E0220 00:11:40.601770 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9641156-dece-4bc9-8734-0b77e320f013" containerName="pruner" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.601776 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9641156-dece-4bc9-8734-0b77e320f013" containerName="pruner" Feb 20 00:11:40 crc kubenswrapper[4750]: E0220 00:11:40.601785 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e2d40b9-179b-473d-b639-6c2ba54f2815" containerName="route-controller-manager" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.601791 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e2d40b9-179b-473d-b639-6c2ba54f2815" containerName="route-controller-manager" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.601878 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce4f9b76-9136-4deb-b4da-7d5c6e5efad7" containerName="pruner" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.601889 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e2d40b9-179b-473d-b639-6c2ba54f2815" containerName="route-controller-manager" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.601897 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9641156-dece-4bc9-8734-0b77e320f013" containerName="pruner" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.602240 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.603439 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.604259 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.604396 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.606016 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.606166 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.606743 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.612040 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm"] Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.645950 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-client-ca\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.646012 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-config\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.647632 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-serving-cert\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.647725 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bh46\" (UniqueName: \"kubernetes.io/projected/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-kube-api-access-7bh46\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.748536 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bh46\" (UniqueName: \"kubernetes.io/projected/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-kube-api-access-7bh46\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.748618 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-client-ca\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.748658 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-config\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.748700 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-serving-cert\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.750207 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-client-ca\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.751816 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-config\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.763440 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-serving-cert\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.764396 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bh46\" (UniqueName: \"kubernetes.io/projected/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-kube-api-access-7bh46\") pod \"route-controller-manager-5f86b664bc-ppqmm\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:40 crc kubenswrapper[4750]: I0220 00:11:40.922937 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:45 crc kubenswrapper[4750]: E0220 00:11:45.173861 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 20 00:11:45 crc kubenswrapper[4750]: I0220 00:11:45.175656 4750 generic.go:334] "Generic (PLEG): container finished" podID="bcd2fc23-8170-4bb9-90a2-5909cbea778a" containerID="2e87b0a66e13d66bc4cd81bad31c0901488a963878574e3d0cbeffee8e20dd04" exitCode=0 Feb 20 00:11:45 crc kubenswrapper[4750]: I0220 00:11:45.175721 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29525760-6k6m7" event={"ID":"bcd2fc23-8170-4bb9-90a2-5909cbea778a","Type":"ContainerDied","Data":"2e87b0a66e13d66bc4cd81bad31c0901488a963878574e3d0cbeffee8e20dd04"} Feb 20 00:11:45 crc kubenswrapper[4750]: E0220 00:11:45.175775 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vklkw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-62qrs_openshift-marketplace(3a63cf32-f756-45bb-ab8d-1e3aff3d2165): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 00:11:45 crc kubenswrapper[4750]: E0220 00:11:45.176994 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-62qrs" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" Feb 20 00:11:48 crc kubenswrapper[4750]: E0220 00:11:48.313187 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-62qrs" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" Feb 20 00:11:48 crc kubenswrapper[4750]: E0220 00:11:48.452882 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 20 00:11:48 crc kubenswrapper[4750]: E0220 00:11:48.453163 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5qqzn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rgp2g_openshift-marketplace(9ce86efc-8081-4418-ac52-06ce3348bc69): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 00:11:48 crc kubenswrapper[4750]: E0220 00:11:48.455659 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rgp2g" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" Feb 20 00:11:49 crc kubenswrapper[4750]: I0220 00:11:49.869868 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-sz9qr" Feb 20 00:11:50 crc kubenswrapper[4750]: I0220 00:11:50.104404 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm"] Feb 20 00:11:50 crc kubenswrapper[4750]: I0220 00:11:50.236698 4750 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9bxjm container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 00:11:50 crc kubenswrapper[4750]: I0220 00:11:50.236759 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.259646 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rgp2g" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.349797 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.349982 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5w5wt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-g8v8d_openshift-marketplace(e21f8aa9-f543-4f68-b241-fde6f0e6bb49): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.351417 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-g8v8d" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.384614 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.384956 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2sjhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-5cq9b_openshift-marketplace(ac3d2b05-2e15-4058-a6f1-1ef965755b02): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.386322 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-5cq9b" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.401612 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.401731 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wlgqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-sll9r_openshift-marketplace(295b160a-bc39-40be-822d-fc4438895480): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.402826 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-sll9r" podUID="295b160a-bc39-40be-822d-fc4438895480" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.413244 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.427648 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.447944 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79df6c4bb4-vtb65"] Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.448170 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.448182 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" Feb 20 00:11:52 crc kubenswrapper[4750]: E0220 00:11:52.448194 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd2fc23-8170-4bb9-90a2-5909cbea778a" containerName="image-pruner" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.448201 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd2fc23-8170-4bb9-90a2-5909cbea778a" containerName="image-pruner" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.448287 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" containerName="controller-manager" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.448300 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd2fc23-8170-4bb9-90a2-5909cbea778a" containerName="image-pruner" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.448639 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.461097 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79df6c4bb4-vtb65"] Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.504096 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bcd2fc23-8170-4bb9-90a2-5909cbea778a-serviceca\") pod \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.504429 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nld4x\" (UniqueName: \"kubernetes.io/projected/98d80369-7e37-4be7-beea-22dc5ef528c3-kube-api-access-nld4x\") pod \"98d80369-7e37-4be7-beea-22dc5ef528c3\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.504898 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcd2fc23-8170-4bb9-90a2-5909cbea778a-serviceca" (OuterVolumeSpecName: "serviceca") pod "bcd2fc23-8170-4bb9-90a2-5909cbea778a" (UID: "bcd2fc23-8170-4bb9-90a2-5909cbea778a"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.505273 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98d80369-7e37-4be7-beea-22dc5ef528c3-serving-cert\") pod \"98d80369-7e37-4be7-beea-22dc5ef528c3\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.505320 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-proxy-ca-bundles\") pod \"98d80369-7e37-4be7-beea-22dc5ef528c3\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.505995 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-client-ca\") pod \"98d80369-7e37-4be7-beea-22dc5ef528c3\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.506024 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-config\") pod \"98d80369-7e37-4be7-beea-22dc5ef528c3\" (UID: \"98d80369-7e37-4be7-beea-22dc5ef528c3\") " Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.506054 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs9sq\" (UniqueName: \"kubernetes.io/projected/bcd2fc23-8170-4bb9-90a2-5909cbea778a-kube-api-access-hs9sq\") pod \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\" (UID: \"bcd2fc23-8170-4bb9-90a2-5909cbea778a\") " Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.506331 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-proxy-ca-bundles\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.505951 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "98d80369-7e37-4be7-beea-22dc5ef528c3" (UID: "98d80369-7e37-4be7-beea-22dc5ef528c3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.506650 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-client-ca" (OuterVolumeSpecName: "client-ca") pod "98d80369-7e37-4be7-beea-22dc5ef528c3" (UID: "98d80369-7e37-4be7-beea-22dc5ef528c3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507041 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-config" (OuterVolumeSpecName: "config") pod "98d80369-7e37-4be7-beea-22dc5ef528c3" (UID: "98d80369-7e37-4be7-beea-22dc5ef528c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507170 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-config\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507208 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-serving-cert\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507236 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pgv5\" (UniqueName: \"kubernetes.io/projected/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-kube-api-access-7pgv5\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507284 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-client-ca\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507325 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507338 4750 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bcd2fc23-8170-4bb9-90a2-5909cbea778a-serviceca\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507349 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.507359 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98d80369-7e37-4be7-beea-22dc5ef528c3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.510768 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d80369-7e37-4be7-beea-22dc5ef528c3-kube-api-access-nld4x" (OuterVolumeSpecName: "kube-api-access-nld4x") pod "98d80369-7e37-4be7-beea-22dc5ef528c3" (UID: "98d80369-7e37-4be7-beea-22dc5ef528c3"). InnerVolumeSpecName "kube-api-access-nld4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.514500 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd2fc23-8170-4bb9-90a2-5909cbea778a-kube-api-access-hs9sq" (OuterVolumeSpecName: "kube-api-access-hs9sq") pod "bcd2fc23-8170-4bb9-90a2-5909cbea778a" (UID: "bcd2fc23-8170-4bb9-90a2-5909cbea778a"). InnerVolumeSpecName "kube-api-access-hs9sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.514605 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d80369-7e37-4be7-beea-22dc5ef528c3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "98d80369-7e37-4be7-beea-22dc5ef528c3" (UID: "98d80369-7e37-4be7-beea-22dc5ef528c3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.594145 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm"] Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608049 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-client-ca\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608248 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-proxy-ca-bundles\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608360 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-config\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608452 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-serving-cert\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608558 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pgv5\" (UniqueName: \"kubernetes.io/projected/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-kube-api-access-7pgv5\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608664 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nld4x\" (UniqueName: \"kubernetes.io/projected/98d80369-7e37-4be7-beea-22dc5ef528c3-kube-api-access-nld4x\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608773 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98d80369-7e37-4be7-beea-22dc5ef528c3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.608858 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs9sq\" (UniqueName: \"kubernetes.io/projected/bcd2fc23-8170-4bb9-90a2-5909cbea778a-kube-api-access-hs9sq\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.609166 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-client-ca\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.610448 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-config\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.611005 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-proxy-ca-bundles\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: W0220 00:11:52.611402 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa0f7945_ecfd_4168_bc10_bb54b42ca8eb.slice/crio-1341ec91d9ca7adc6c3a83ee14ffab32c7c259299efb1c68689e70e7d3a16d4b WatchSource:0}: Error finding container 1341ec91d9ca7adc6c3a83ee14ffab32c7c259299efb1c68689e70e7d3a16d4b: Status 404 returned error can't find the container with id 1341ec91d9ca7adc6c3a83ee14ffab32c7c259299efb1c68689e70e7d3a16d4b Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.615047 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-serving-cert\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.629486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pgv5\" (UniqueName: \"kubernetes.io/projected/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-kube-api-access-7pgv5\") pod \"controller-manager-79df6c4bb4-vtb65\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.681976 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tcgh4"] Feb 20 00:11:52 crc kubenswrapper[4750]: W0220 00:11:52.709076 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfac6ee29_99b8_42e7_844d_30e68cbecad3.slice/crio-dcdba01ed671260724f4235a24a3a705493f6a56977b942baa8295191a55677a WatchSource:0}: Error finding container dcdba01ed671260724f4235a24a3a705493f6a56977b942baa8295191a55677a: Status 404 returned error can't find the container with id dcdba01ed671260724f4235a24a3a705493f6a56977b942baa8295191a55677a Feb 20 00:11:52 crc kubenswrapper[4750]: I0220 00:11:52.820908 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.027933 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79df6c4bb4-vtb65"] Feb 20 00:11:53 crc kubenswrapper[4750]: W0220 00:11:53.045727 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92e9fbea_27b0_4b5f_a4d8_6a014836dad1.slice/crio-e18e65dd30f66a1c82300289a30c6df893fdcd06342ce806bcfd5868f94566b7 WatchSource:0}: Error finding container e18e65dd30f66a1c82300289a30c6df893fdcd06342ce806bcfd5868f94566b7: Status 404 returned error can't find the container with id e18e65dd30f66a1c82300289a30c6df893fdcd06342ce806bcfd5868f94566b7 Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.252040 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" event={"ID":"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb","Type":"ContainerStarted","Data":"83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.252430 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.252441 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" event={"ID":"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb","Type":"ContainerStarted","Data":"1341ec91d9ca7adc6c3a83ee14ffab32c7c259299efb1c68689e70e7d3a16d4b"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.252137 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" podUID="aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" containerName="route-controller-manager" containerID="cri-o://83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c" gracePeriod=30 Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.255587 4750 generic.go:334] "Generic (PLEG): container finished" podID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerID="1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285" exitCode=0 Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.255653 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjlwq" event={"ID":"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7","Type":"ContainerDied","Data":"1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.259480 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.260316 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" event={"ID":"92e9fbea-27b0-4b5f-a4d8-6a014836dad1","Type":"ContainerStarted","Data":"b2b3629b491bea280dcc67153335ea208031b2451293cbb046bc0127c2005dff"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.260369 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" event={"ID":"92e9fbea-27b0-4b5f-a4d8-6a014836dad1","Type":"ContainerStarted","Data":"e18e65dd30f66a1c82300289a30c6df893fdcd06342ce806bcfd5868f94566b7"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.261147 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.264551 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" event={"ID":"fac6ee29-99b8-42e7-844d-30e68cbecad3","Type":"ContainerStarted","Data":"515ca5df05d8ad67f577664a651cf6d7c552b652b555ce1a4ef242c45464bc7d"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.264607 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" event={"ID":"fac6ee29-99b8-42e7-844d-30e68cbecad3","Type":"ContainerStarted","Data":"d799ba90485824733162f7cbe787ecb3c1f558e5adfdaac8ce036511829b0260"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.264623 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tcgh4" event={"ID":"fac6ee29-99b8-42e7-844d-30e68cbecad3","Type":"ContainerStarted","Data":"dcdba01ed671260724f4235a24a3a705493f6a56977b942baa8295191a55677a"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.267028 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29525760-6k6m7" event={"ID":"bcd2fc23-8170-4bb9-90a2-5909cbea778a","Type":"ContainerDied","Data":"65ad14a9a75f0ae998193984f441ec064072e28bdc05228cb6db7d98dc905ae3"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.267062 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65ad14a9a75f0ae998193984f441ec064072e28bdc05228cb6db7d98dc905ae3" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.267148 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29525760-6k6m7" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.267826 4750 patch_prober.go:28] interesting pod/controller-manager-79df6c4bb4-vtb65 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.267855 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" podUID="92e9fbea-27b0-4b5f-a4d8-6a014836dad1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.269150 4750 generic.go:334] "Generic (PLEG): container finished" podID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerID="84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0" exitCode=0 Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.269200 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbjg" event={"ID":"9595c598-a6fe-45fc-a934-2a64ca8ef288","Type":"ContainerDied","Data":"84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.275016 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" event={"ID":"98d80369-7e37-4be7-beea-22dc5ef528c3","Type":"ContainerDied","Data":"2ae02951e720a8a425fc90b7f8ed83063b5f6eea4963d6c89c7649e3c01d5773"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.275067 4750 scope.go:117] "RemoveContainer" containerID="cc27c4ae5466e9f2596e970bde1cf3d1e16e40223aab82c1679451087e27783f" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.275033 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9bxjm" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.277787 4750 generic.go:334] "Generic (PLEG): container finished" podID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerID="de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698" exitCode=0 Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.279178 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2vd9" event={"ID":"0e3cf328-65f5-417b-8bb4-35ff194e8cc4","Type":"ContainerDied","Data":"de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698"} Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.280064 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" podStartSLOduration=23.280048809 podStartE2EDuration="23.280048809s" podCreationTimestamp="2026-02-20 00:11:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:53.276397439 +0000 UTC m=+197.471233488" watchObservedRunningTime="2026-02-20 00:11:53.280048809 +0000 UTC m=+197.474884858" Feb 20 00:11:53 crc kubenswrapper[4750]: E0220 00:11:53.280102 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-g8v8d" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" Feb 20 00:11:53 crc kubenswrapper[4750]: E0220 00:11:53.280358 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-sll9r" podUID="295b160a-bc39-40be-822d-fc4438895480" Feb 20 00:11:53 crc kubenswrapper[4750]: E0220 00:11:53.280654 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-5cq9b" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.299562 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-tcgh4" podStartSLOduration=172.299542033 podStartE2EDuration="2m52.299542033s" podCreationTimestamp="2026-02-20 00:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:53.299325837 +0000 UTC m=+197.494161886" watchObservedRunningTime="2026-02-20 00:11:53.299542033 +0000 UTC m=+197.494378082" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.338741 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" podStartSLOduration=3.338725995 podStartE2EDuration="3.338725995s" podCreationTimestamp="2026-02-20 00:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:53.336555036 +0000 UTC m=+197.531391095" watchObservedRunningTime="2026-02-20 00:11:53.338725995 +0000 UTC m=+197.533562034" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.441009 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9bxjm"] Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.446531 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9bxjm"] Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.735639 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.826939 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-serving-cert\") pod \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.827326 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-client-ca\") pod \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.827360 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bh46\" (UniqueName: \"kubernetes.io/projected/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-kube-api-access-7bh46\") pod \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.827398 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-config\") pod \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\" (UID: \"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb\") " Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.828131 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-config" (OuterVolumeSpecName: "config") pod "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" (UID: "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.828313 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-client-ca" (OuterVolumeSpecName: "client-ca") pod "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" (UID: "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.834656 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-kube-api-access-7bh46" (OuterVolumeSpecName: "kube-api-access-7bh46") pod "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" (UID: "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb"). InnerVolumeSpecName "kube-api-access-7bh46". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.835223 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" (UID: "aa0f7945-ecfd-4168-bc10-bb54b42ca8eb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.928825 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.928860 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bh46\" (UniqueName: \"kubernetes.io/projected/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-kube-api-access-7bh46\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.928870 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:53 crc kubenswrapper[4750]: I0220 00:11:53.928880 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.288307 4750 generic.go:334] "Generic (PLEG): container finished" podID="aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" containerID="83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c" exitCode=0 Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.288393 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" event={"ID":"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb","Type":"ContainerDied","Data":"83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c"} Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.288431 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" event={"ID":"aa0f7945-ecfd-4168-bc10-bb54b42ca8eb","Type":"ContainerDied","Data":"1341ec91d9ca7adc6c3a83ee14ffab32c7c259299efb1c68689e70e7d3a16d4b"} Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.288457 4750 scope.go:117] "RemoveContainer" containerID="83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.288558 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.295685 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjlwq" event={"ID":"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7","Type":"ContainerStarted","Data":"f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5"} Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.306600 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2vd9" event={"ID":"0e3cf328-65f5-417b-8bb4-35ff194e8cc4","Type":"ContainerStarted","Data":"67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b"} Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.309675 4750 scope.go:117] "RemoveContainer" containerID="83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c" Feb 20 00:11:54 crc kubenswrapper[4750]: E0220 00:11:54.310113 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c\": container with ID starting with 83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c not found: ID does not exist" containerID="83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.310194 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c"} err="failed to get container status \"83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c\": rpc error: code = NotFound desc = could not find container \"83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c\": container with ID starting with 83de180212aca774e093c9a1a2012549ca9c75b87b7f5ea3c5ffa818e85bfa3c not found: ID does not exist" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.313285 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbjg" event={"ID":"9595c598-a6fe-45fc-a934-2a64ca8ef288","Type":"ContainerStarted","Data":"de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44"} Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.322537 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.347865 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gjlwq" podStartSLOduration=2.315991957 podStartE2EDuration="38.347842232s" podCreationTimestamp="2026-02-20 00:11:16 +0000 UTC" firstStartedPulling="2026-02-20 00:11:17.682320319 +0000 UTC m=+161.877156368" lastFinishedPulling="2026-02-20 00:11:53.714170594 +0000 UTC m=+197.909006643" observedRunningTime="2026-02-20 00:11:54.322005805 +0000 UTC m=+198.516841884" watchObservedRunningTime="2026-02-20 00:11:54.347842232 +0000 UTC m=+198.542678301" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.351415 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4zbjg" podStartSLOduration=3.4674437129999998 podStartE2EDuration="36.35140257s" podCreationTimestamp="2026-02-20 00:11:18 +0000 UTC" firstStartedPulling="2026-02-20 00:11:20.806863748 +0000 UTC m=+165.001699797" lastFinishedPulling="2026-02-20 00:11:53.690822595 +0000 UTC m=+197.885658654" observedRunningTime="2026-02-20 00:11:54.342742123 +0000 UTC m=+198.537578222" watchObservedRunningTime="2026-02-20 00:11:54.35140257 +0000 UTC m=+198.546238629" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.363651 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm"] Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.369377 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f86b664bc-ppqmm"] Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.388612 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z2vd9" podStartSLOduration=2.245882573 podStartE2EDuration="36.388589978s" podCreationTimestamp="2026-02-20 00:11:18 +0000 UTC" firstStartedPulling="2026-02-20 00:11:19.75632131 +0000 UTC m=+163.951157349" lastFinishedPulling="2026-02-20 00:11:53.899028705 +0000 UTC m=+198.093864754" observedRunningTime="2026-02-20 00:11:54.387199859 +0000 UTC m=+198.582035938" watchObservedRunningTime="2026-02-20 00:11:54.388589978 +0000 UTC m=+198.583426037" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.567826 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98d80369-7e37-4be7-beea-22dc5ef528c3" path="/var/lib/kubelet/pods/98d80369-7e37-4be7-beea-22dc5ef528c3/volumes" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.568696 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" path="/var/lib/kubelet/pods/aa0f7945-ecfd-4168-bc10-bb54b42ca8eb/volumes" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.622726 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv"] Feb 20 00:11:54 crc kubenswrapper[4750]: E0220 00:11:54.623037 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" containerName="route-controller-manager" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.623065 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" containerName="route-controller-manager" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.623292 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0f7945-ecfd-4168-bc10-bb54b42ca8eb" containerName="route-controller-manager" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.623878 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.626430 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.627303 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.627953 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.627307 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.629432 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.636408 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb740d0-8b1d-4e74-980c-417f9b4cc855-serving-cert\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.636451 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-client-ca\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.636488 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-config\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.636514 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m498z\" (UniqueName: \"kubernetes.io/projected/9fb740d0-8b1d-4e74-980c-417f9b4cc855-kube-api-access-m498z\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.636630 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv"] Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.637100 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.737598 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m498z\" (UniqueName: \"kubernetes.io/projected/9fb740d0-8b1d-4e74-980c-417f9b4cc855-kube-api-access-m498z\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.737709 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb740d0-8b1d-4e74-980c-417f9b4cc855-serving-cert\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.737743 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-client-ca\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.737788 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-config\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.738669 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-client-ca\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.738983 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-config\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.749989 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb740d0-8b1d-4e74-980c-417f9b4cc855-serving-cert\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.752722 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m498z\" (UniqueName: \"kubernetes.io/projected/9fb740d0-8b1d-4e74-980c-417f9b4cc855-kube-api-access-m498z\") pod \"route-controller-manager-6d5b67bccc-f26qv\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:54 crc kubenswrapper[4750]: I0220 00:11:54.938151 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:55 crc kubenswrapper[4750]: I0220 00:11:55.368596 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv"] Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.327027 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" event={"ID":"9fb740d0-8b1d-4e74-980c-417f9b4cc855","Type":"ContainerStarted","Data":"f34ca3254e5e46aee4bddda2fbb4b53cc03a000333b4add97aa0cdb7e64dcbf9"} Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.327309 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" event={"ID":"9fb740d0-8b1d-4e74-980c-417f9b4cc855","Type":"ContainerStarted","Data":"0e41b0378b95b13c12080cb41cffa68a71f5ac90f216e8719f3a7da8b2ec4900"} Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.341343 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" podStartSLOduration=6.341326187 podStartE2EDuration="6.341326187s" podCreationTimestamp="2026-02-20 00:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:11:56.340586487 +0000 UTC m=+200.535422536" watchObservedRunningTime="2026-02-20 00:11:56.341326187 +0000 UTC m=+200.536162226" Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.669673 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.669754 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.733136 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.733193 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:56 crc kubenswrapper[4750]: I0220 00:11:56.918033 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:11:57 crc kubenswrapper[4750]: I0220 00:11:57.339710 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:57 crc kubenswrapper[4750]: I0220 00:11:57.346496 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:11:57 crc kubenswrapper[4750]: I0220 00:11:57.806177 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nd2j6"] Feb 20 00:11:58 crc kubenswrapper[4750]: I0220 00:11:58.574170 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:58 crc kubenswrapper[4750]: I0220 00:11:58.574205 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:58 crc kubenswrapper[4750]: I0220 00:11:58.608811 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:58 crc kubenswrapper[4750]: I0220 00:11:58.936950 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:58 crc kubenswrapper[4750]: I0220 00:11:58.937418 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.002005 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.027893 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.028502 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.035677 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.038384 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.038612 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.167588 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a22927d9-5852-4701-9cad-4e78becdb8f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.167925 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a22927d9-5852-4701-9cad-4e78becdb8f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.268680 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a22927d9-5852-4701-9cad-4e78becdb8f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.268741 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a22927d9-5852-4701-9cad-4e78becdb8f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.268830 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a22927d9-5852-4701-9cad-4e78becdb8f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.291734 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a22927d9-5852-4701-9cad-4e78becdb8f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.359676 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.392275 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.394265 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:11:59 crc kubenswrapper[4750]: I0220 00:11:59.786707 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 00:12:00 crc kubenswrapper[4750]: I0220 00:12:00.729495 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a22927d9-5852-4701-9cad-4e78becdb8f6","Type":"ContainerStarted","Data":"d5860e6948db4f0a3d1b5910aecc66db4dd95fb857b687a4fbc75a804d37262d"} Feb 20 00:12:00 crc kubenswrapper[4750]: I0220 00:12:00.729867 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a22927d9-5852-4701-9cad-4e78becdb8f6","Type":"ContainerStarted","Data":"fb196133a5758cf8746d6949e07b56b0460e6e30bae0746418433cf830300828"} Feb 20 00:12:00 crc kubenswrapper[4750]: I0220 00:12:00.746770 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.746753145 podStartE2EDuration="1.746753145s" podCreationTimestamp="2026-02-20 00:11:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:12:00.744315908 +0000 UTC m=+204.939151957" watchObservedRunningTime="2026-02-20 00:12:00.746753145 +0000 UTC m=+204.941589194" Feb 20 00:12:01 crc kubenswrapper[4750]: I0220 00:12:01.734411 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62qrs" event={"ID":"3a63cf32-f756-45bb-ab8d-1e3aff3d2165","Type":"ContainerStarted","Data":"091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733"} Feb 20 00:12:01 crc kubenswrapper[4750]: I0220 00:12:01.734999 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbjg"] Feb 20 00:12:01 crc kubenswrapper[4750]: I0220 00:12:01.735277 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4zbjg" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="registry-server" containerID="cri-o://de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44" gracePeriod=2 Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.677995 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.741197 4750 generic.go:334] "Generic (PLEG): container finished" podID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerID="de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44" exitCode=0 Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.741264 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4zbjg" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.741292 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbjg" event={"ID":"9595c598-a6fe-45fc-a934-2a64ca8ef288","Type":"ContainerDied","Data":"de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44"} Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.741335 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4zbjg" event={"ID":"9595c598-a6fe-45fc-a934-2a64ca8ef288","Type":"ContainerDied","Data":"b71a858dba8de9842ed96380aeab2a9eaecb2a4d928e7d44c1cecebc4a26e3f8"} Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.741358 4750 scope.go:117] "RemoveContainer" containerID="de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.744087 4750 generic.go:334] "Generic (PLEG): container finished" podID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerID="091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733" exitCode=0 Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.744144 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62qrs" event={"ID":"3a63cf32-f756-45bb-ab8d-1e3aff3d2165","Type":"ContainerDied","Data":"091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733"} Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.760647 4750 scope.go:117] "RemoveContainer" containerID="84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.780902 4750 scope.go:117] "RemoveContainer" containerID="f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.815489 4750 scope.go:117] "RemoveContainer" containerID="de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44" Feb 20 00:12:02 crc kubenswrapper[4750]: E0220 00:12:02.815943 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44\": container with ID starting with de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44 not found: ID does not exist" containerID="de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.815984 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44"} err="failed to get container status \"de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44\": rpc error: code = NotFound desc = could not find container \"de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44\": container with ID starting with de6deeb3779a563258be20c8da66a41ad8cd40a7bb4e59c3f9ff24c1c59e8c44 not found: ID does not exist" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.816011 4750 scope.go:117] "RemoveContainer" containerID="84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0" Feb 20 00:12:02 crc kubenswrapper[4750]: E0220 00:12:02.816367 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0\": container with ID starting with 84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0 not found: ID does not exist" containerID="84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.816399 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0"} err="failed to get container status \"84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0\": rpc error: code = NotFound desc = could not find container \"84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0\": container with ID starting with 84ed8e5bbc159f88f335565f98768fcfbf4de0215932ec8b0561423c9eefa1f0 not found: ID does not exist" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.816426 4750 scope.go:117] "RemoveContainer" containerID="f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f" Feb 20 00:12:02 crc kubenswrapper[4750]: E0220 00:12:02.816684 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f\": container with ID starting with f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f not found: ID does not exist" containerID="f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.816708 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f"} err="failed to get container status \"f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f\": rpc error: code = NotFound desc = could not find container \"f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f\": container with ID starting with f4e212f4a7822a839261eb11e56d768086871858f282f32073386bdfbe1df40f not found: ID does not exist" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.868092 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-catalog-content\") pod \"9595c598-a6fe-45fc-a934-2a64ca8ef288\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.868158 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjqm9\" (UniqueName: \"kubernetes.io/projected/9595c598-a6fe-45fc-a934-2a64ca8ef288-kube-api-access-kjqm9\") pod \"9595c598-a6fe-45fc-a934-2a64ca8ef288\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.868219 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-utilities\") pod \"9595c598-a6fe-45fc-a934-2a64ca8ef288\" (UID: \"9595c598-a6fe-45fc-a934-2a64ca8ef288\") " Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.869464 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-utilities" (OuterVolumeSpecName: "utilities") pod "9595c598-a6fe-45fc-a934-2a64ca8ef288" (UID: "9595c598-a6fe-45fc-a934-2a64ca8ef288"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.879240 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9595c598-a6fe-45fc-a934-2a64ca8ef288-kube-api-access-kjqm9" (OuterVolumeSpecName: "kube-api-access-kjqm9") pod "9595c598-a6fe-45fc-a934-2a64ca8ef288" (UID: "9595c598-a6fe-45fc-a934-2a64ca8ef288"). InnerVolumeSpecName "kube-api-access-kjqm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.901894 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9595c598-a6fe-45fc-a934-2a64ca8ef288" (UID: "9595c598-a6fe-45fc-a934-2a64ca8ef288"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.969568 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.969607 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjqm9\" (UniqueName: \"kubernetes.io/projected/9595c598-a6fe-45fc-a934-2a64ca8ef288-kube-api-access-kjqm9\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:02 crc kubenswrapper[4750]: I0220 00:12:02.969620 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9595c598-a6fe-45fc-a934-2a64ca8ef288-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:03 crc kubenswrapper[4750]: I0220 00:12:03.065033 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbjg"] Feb 20 00:12:03 crc kubenswrapper[4750]: I0220 00:12:03.068697 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4zbjg"] Feb 20 00:12:03 crc kubenswrapper[4750]: I0220 00:12:03.752487 4750 generic.go:334] "Generic (PLEG): container finished" podID="a22927d9-5852-4701-9cad-4e78becdb8f6" containerID="d5860e6948db4f0a3d1b5910aecc66db4dd95fb857b687a4fbc75a804d37262d" exitCode=0 Feb 20 00:12:03 crc kubenswrapper[4750]: I0220 00:12:03.752577 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a22927d9-5852-4701-9cad-4e78becdb8f6","Type":"ContainerDied","Data":"d5860e6948db4f0a3d1b5910aecc66db4dd95fb857b687a4fbc75a804d37262d"} Feb 20 00:12:03 crc kubenswrapper[4750]: I0220 00:12:03.760293 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62qrs" event={"ID":"3a63cf32-f756-45bb-ab8d-1e3aff3d2165","Type":"ContainerStarted","Data":"03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a"} Feb 20 00:12:03 crc kubenswrapper[4750]: I0220 00:12:03.790408 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-62qrs" podStartSLOduration=2.363318421 podStartE2EDuration="47.79038826s" podCreationTimestamp="2026-02-20 00:11:16 +0000 UTC" firstStartedPulling="2026-02-20 00:11:17.677774923 +0000 UTC m=+161.872610972" lastFinishedPulling="2026-02-20 00:12:03.104844762 +0000 UTC m=+207.299680811" observedRunningTime="2026-02-20 00:12:03.786640528 +0000 UTC m=+207.981476597" watchObservedRunningTime="2026-02-20 00:12:03.79038826 +0000 UTC m=+207.985224309" Feb 20 00:12:04 crc kubenswrapper[4750]: I0220 00:12:04.569919 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" path="/var/lib/kubelet/pods/9595c598-a6fe-45fc-a934-2a64ca8ef288/volumes" Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.248999 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.400672 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a22927d9-5852-4701-9cad-4e78becdb8f6-kube-api-access\") pod \"a22927d9-5852-4701-9cad-4e78becdb8f6\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.400805 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a22927d9-5852-4701-9cad-4e78becdb8f6-kubelet-dir\") pod \"a22927d9-5852-4701-9cad-4e78becdb8f6\" (UID: \"a22927d9-5852-4701-9cad-4e78becdb8f6\") " Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.400929 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a22927d9-5852-4701-9cad-4e78becdb8f6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a22927d9-5852-4701-9cad-4e78becdb8f6" (UID: "a22927d9-5852-4701-9cad-4e78becdb8f6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.401281 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a22927d9-5852-4701-9cad-4e78becdb8f6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.406360 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a22927d9-5852-4701-9cad-4e78becdb8f6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a22927d9-5852-4701-9cad-4e78becdb8f6" (UID: "a22927d9-5852-4701-9cad-4e78becdb8f6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.502444 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a22927d9-5852-4701-9cad-4e78becdb8f6-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.772266 4750 generic.go:334] "Generic (PLEG): container finished" podID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerID="55250ea9abcac1cf7012db718d4261da7c430216e4687651e136ca16dea6bb7a" exitCode=0 Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.772337 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgp2g" event={"ID":"9ce86efc-8081-4418-ac52-06ce3348bc69","Type":"ContainerDied","Data":"55250ea9abcac1cf7012db718d4261da7c430216e4687651e136ca16dea6bb7a"} Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.774060 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a22927d9-5852-4701-9cad-4e78becdb8f6","Type":"ContainerDied","Data":"fb196133a5758cf8746d6949e07b56b0460e6e30bae0746418433cf830300828"} Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.774184 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 00:12:05 crc kubenswrapper[4750]: I0220 00:12:05.774109 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb196133a5758cf8746d6949e07b56b0460e6e30bae0746418433cf830300828" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.603671 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.606093 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.634651 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 00:12:06 crc kubenswrapper[4750]: E0220 00:12:06.634930 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22927d9-5852-4701-9cad-4e78becdb8f6" containerName="pruner" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.634951 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22927d9-5852-4701-9cad-4e78becdb8f6" containerName="pruner" Feb 20 00:12:06 crc kubenswrapper[4750]: E0220 00:12:06.634969 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="extract-utilities" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.634977 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="extract-utilities" Feb 20 00:12:06 crc kubenswrapper[4750]: E0220 00:12:06.634993 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="extract-content" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.635002 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="extract-content" Feb 20 00:12:06 crc kubenswrapper[4750]: E0220 00:12:06.635012 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="registry-server" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.635021 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="registry-server" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.635144 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22927d9-5852-4701-9cad-4e78becdb8f6" containerName="pruner" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.635161 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9595c598-a6fe-45fc-a934-2a64ca8ef288" containerName="registry-server" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.636649 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.638494 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.638841 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.669991 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.676400 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.769101 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.817086 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.817206 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kube-api-access\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.817283 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-var-lock\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.918236 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-var-lock\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.918345 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-var-lock\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.918450 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.918497 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.918520 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kube-api-access\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.939404 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kube-api-access\") pod \"installer-9-crc\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:06 crc kubenswrapper[4750]: I0220 00:12:06.962426 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:07 crc kubenswrapper[4750]: I0220 00:12:07.435493 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 00:12:07 crc kubenswrapper[4750]: W0220 00:12:07.445210 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbf2cf62d_cf5c_429a_b43c_228bae403a6e.slice/crio-6fa86484ac1fc81e50a71dced032b53e165c2ee36fa00e7716d80752e9520795 WatchSource:0}: Error finding container 6fa86484ac1fc81e50a71dced032b53e165c2ee36fa00e7716d80752e9520795: Status 404 returned error can't find the container with id 6fa86484ac1fc81e50a71dced032b53e165c2ee36fa00e7716d80752e9520795 Feb 20 00:12:07 crc kubenswrapper[4750]: I0220 00:12:07.802948 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgp2g" event={"ID":"9ce86efc-8081-4418-ac52-06ce3348bc69","Type":"ContainerStarted","Data":"c2649448706f267d1e7a8cc5fef1d582664d265352c8ffc1ad9ec8cd50e43b3e"} Feb 20 00:12:07 crc kubenswrapper[4750]: I0220 00:12:07.808434 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bf2cf62d-cf5c-429a-b43c-228bae403a6e","Type":"ContainerStarted","Data":"6a3358f6e2340985b1d720644f8d3eabae89cca193a0f8786515bba5b41be3c9"} Feb 20 00:12:07 crc kubenswrapper[4750]: I0220 00:12:07.808560 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bf2cf62d-cf5c-429a-b43c-228bae403a6e","Type":"ContainerStarted","Data":"6fa86484ac1fc81e50a71dced032b53e165c2ee36fa00e7716d80752e9520795"} Feb 20 00:12:07 crc kubenswrapper[4750]: I0220 00:12:07.844241 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.844220942 podStartE2EDuration="1.844220942s" podCreationTimestamp="2026-02-20 00:12:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:12:07.842357281 +0000 UTC m=+212.037193350" watchObservedRunningTime="2026-02-20 00:12:07.844220942 +0000 UTC m=+212.039056991" Feb 20 00:12:07 crc kubenswrapper[4750]: I0220 00:12:07.845433 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rgp2g" podStartSLOduration=2.89681161 podStartE2EDuration="51.845426875s" podCreationTimestamp="2026-02-20 00:11:16 +0000 UTC" firstStartedPulling="2026-02-20 00:11:17.696993822 +0000 UTC m=+161.891829871" lastFinishedPulling="2026-02-20 00:12:06.645609077 +0000 UTC m=+210.840445136" observedRunningTime="2026-02-20 00:12:07.826335132 +0000 UTC m=+212.021171191" watchObservedRunningTime="2026-02-20 00:12:07.845426875 +0000 UTC m=+212.040262924" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.133536 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gjlwq"] Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.133753 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gjlwq" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="registry-server" containerID="cri-o://f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5" gracePeriod=2 Feb 20 00:12:08 crc kubenswrapper[4750]: E0220 00:12:08.224421 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac3d2b05_2e15_4058_a6f1_1ef965755b02.slice/crio-921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac3d2b05_2e15_4058_a6f1_1ef965755b02.slice/crio-conmon-921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886.scope\": RecentStats: unable to find data in memory cache]" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.602729 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.644228 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-catalog-content\") pod \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.644328 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-utilities\") pod \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.644414 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xqh5\" (UniqueName: \"kubernetes.io/projected/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-kube-api-access-8xqh5\") pod \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\" (UID: \"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7\") " Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.648392 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-utilities" (OuterVolumeSpecName: "utilities") pod "7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" (UID: "7fd8558f-eb0c-4f43-aa1f-809ec986a7a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.655725 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-kube-api-access-8xqh5" (OuterVolumeSpecName: "kube-api-access-8xqh5") pod "7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" (UID: "7fd8558f-eb0c-4f43-aa1f-809ec986a7a7"). InnerVolumeSpecName "kube-api-access-8xqh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.722158 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" (UID: "7fd8558f-eb0c-4f43-aa1f-809ec986a7a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.746034 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.746072 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.746082 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xqh5\" (UniqueName: \"kubernetes.io/projected/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7-kube-api-access-8xqh5\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.817894 4750 generic.go:334] "Generic (PLEG): container finished" podID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerID="78a18902442e10656e64a818e69f07e80541f35e124c18336c58b9dd36e1dec1" exitCode=0 Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.817976 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8v8d" event={"ID":"e21f8aa9-f543-4f68-b241-fde6f0e6bb49","Type":"ContainerDied","Data":"78a18902442e10656e64a818e69f07e80541f35e124c18336c58b9dd36e1dec1"} Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.827476 4750 generic.go:334] "Generic (PLEG): container finished" podID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerID="921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886" exitCode=0 Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.827679 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5cq9b" event={"ID":"ac3d2b05-2e15-4058-a6f1-1ef965755b02","Type":"ContainerDied","Data":"921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886"} Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.831206 4750 generic.go:334] "Generic (PLEG): container finished" podID="295b160a-bc39-40be-822d-fc4438895480" containerID="b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93" exitCode=0 Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.831312 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sll9r" event={"ID":"295b160a-bc39-40be-822d-fc4438895480","Type":"ContainerDied","Data":"b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93"} Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.836034 4750 generic.go:334] "Generic (PLEG): container finished" podID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerID="f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5" exitCode=0 Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.836153 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjlwq" event={"ID":"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7","Type":"ContainerDied","Data":"f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5"} Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.836198 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjlwq" event={"ID":"7fd8558f-eb0c-4f43-aa1f-809ec986a7a7","Type":"ContainerDied","Data":"eb7daec71d0b615c7df4c4eeee250fbec344005cbc561c3ab947529d5163699b"} Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.836229 4750 scope.go:117] "RemoveContainer" containerID="f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.836386 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjlwq" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.854342 4750 scope.go:117] "RemoveContainer" containerID="1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.880964 4750 scope.go:117] "RemoveContainer" containerID="c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.912132 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gjlwq"] Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.914357 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gjlwq"] Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.916282 4750 scope.go:117] "RemoveContainer" containerID="f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5" Feb 20 00:12:08 crc kubenswrapper[4750]: E0220 00:12:08.916742 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5\": container with ID starting with f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5 not found: ID does not exist" containerID="f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.916863 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5"} err="failed to get container status \"f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5\": rpc error: code = NotFound desc = could not find container \"f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5\": container with ID starting with f698e2cb11a5532bf7d2751fe1488043a6fce977c8db3a08e23ff7bc47b437a5 not found: ID does not exist" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.916992 4750 scope.go:117] "RemoveContainer" containerID="1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285" Feb 20 00:12:08 crc kubenswrapper[4750]: E0220 00:12:08.917574 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285\": container with ID starting with 1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285 not found: ID does not exist" containerID="1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.917669 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285"} err="failed to get container status \"1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285\": rpc error: code = NotFound desc = could not find container \"1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285\": container with ID starting with 1db5635e1a0887e4965fbe2445c1e978e235860a9460ca13b836b6f8d918a285 not found: ID does not exist" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.917757 4750 scope.go:117] "RemoveContainer" containerID="c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db" Feb 20 00:12:08 crc kubenswrapper[4750]: E0220 00:12:08.918089 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db\": container with ID starting with c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db not found: ID does not exist" containerID="c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db" Feb 20 00:12:08 crc kubenswrapper[4750]: I0220 00:12:08.918131 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db"} err="failed to get container status \"c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db\": rpc error: code = NotFound desc = could not find container \"c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db\": container with ID starting with c6d7f3601bfde20c88965bda193f1a39b951f0eb41504fb4bacd7770b476a3db not found: ID does not exist" Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.842613 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8v8d" event={"ID":"e21f8aa9-f543-4f68-b241-fde6f0e6bb49","Type":"ContainerStarted","Data":"7706116e03096c00aeeb051fcc8954f37e220c90098d9478e4ee5b3ef814f88a"} Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.845292 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5cq9b" event={"ID":"ac3d2b05-2e15-4058-a6f1-1ef965755b02","Type":"ContainerStarted","Data":"6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5"} Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.847314 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sll9r" event={"ID":"295b160a-bc39-40be-822d-fc4438895480","Type":"ContainerStarted","Data":"d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3"} Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.867088 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8v8d" podStartSLOduration=2.466276611 podStartE2EDuration="50.867070911s" podCreationTimestamp="2026-02-20 00:11:19 +0000 UTC" firstStartedPulling="2026-02-20 00:11:20.820106802 +0000 UTC m=+165.014942851" lastFinishedPulling="2026-02-20 00:12:09.220901102 +0000 UTC m=+213.415737151" observedRunningTime="2026-02-20 00:12:09.863884114 +0000 UTC m=+214.058720183" watchObservedRunningTime="2026-02-20 00:12:09.867070911 +0000 UTC m=+214.061906960" Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.884593 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5cq9b" podStartSLOduration=2.329570319 podStartE2EDuration="53.884573011s" podCreationTimestamp="2026-02-20 00:11:16 +0000 UTC" firstStartedPulling="2026-02-20 00:11:17.686170334 +0000 UTC m=+161.881006383" lastFinishedPulling="2026-02-20 00:12:09.241173026 +0000 UTC m=+213.436009075" observedRunningTime="2026-02-20 00:12:09.883773569 +0000 UTC m=+214.078609628" watchObservedRunningTime="2026-02-20 00:12:09.884573011 +0000 UTC m=+214.079409070" Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.907624 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sll9r" podStartSLOduration=2.418740821 podStartE2EDuration="50.907608642s" podCreationTimestamp="2026-02-20 00:11:19 +0000 UTC" firstStartedPulling="2026-02-20 00:11:20.80728394 +0000 UTC m=+165.002119989" lastFinishedPulling="2026-02-20 00:12:09.296151751 +0000 UTC m=+213.490987810" observedRunningTime="2026-02-20 00:12:09.907308323 +0000 UTC m=+214.102144372" watchObservedRunningTime="2026-02-20 00:12:09.907608642 +0000 UTC m=+214.102444701" Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.978220 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.978584 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.983393 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79df6c4bb4-vtb65"] Feb 20 00:12:09 crc kubenswrapper[4750]: I0220 00:12:09.983615 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" podUID="92e9fbea-27b0-4b5f-a4d8-6a014836dad1" containerName="controller-manager" containerID="cri-o://b2b3629b491bea280dcc67153335ea208031b2451293cbb046bc0127c2005dff" gracePeriod=30 Feb 20 00:12:10 crc kubenswrapper[4750]: I0220 00:12:10.010753 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv"] Feb 20 00:12:10 crc kubenswrapper[4750]: I0220 00:12:10.011163 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" podUID="9fb740d0-8b1d-4e74-980c-417f9b4cc855" containerName="route-controller-manager" containerID="cri-o://f34ca3254e5e46aee4bddda2fbb4b53cc03a000333b4add97aa0cdb7e64dcbf9" gracePeriod=30 Feb 20 00:12:10 crc kubenswrapper[4750]: I0220 00:12:10.565974 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" path="/var/lib/kubelet/pods/7fd8558f-eb0c-4f43-aa1f-809ec986a7a7/volumes" Feb 20 00:12:10 crc kubenswrapper[4750]: I0220 00:12:10.860474 4750 generic.go:334] "Generic (PLEG): container finished" podID="92e9fbea-27b0-4b5f-a4d8-6a014836dad1" containerID="b2b3629b491bea280dcc67153335ea208031b2451293cbb046bc0127c2005dff" exitCode=0 Feb 20 00:12:10 crc kubenswrapper[4750]: I0220 00:12:10.860615 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" event={"ID":"92e9fbea-27b0-4b5f-a4d8-6a014836dad1","Type":"ContainerDied","Data":"b2b3629b491bea280dcc67153335ea208031b2451293cbb046bc0127c2005dff"} Feb 20 00:12:10 crc kubenswrapper[4750]: I0220 00:12:10.865741 4750 generic.go:334] "Generic (PLEG): container finished" podID="9fb740d0-8b1d-4e74-980c-417f9b4cc855" containerID="f34ca3254e5e46aee4bddda2fbb4b53cc03a000333b4add97aa0cdb7e64dcbf9" exitCode=0 Feb 20 00:12:10 crc kubenswrapper[4750]: I0220 00:12:10.865756 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" event={"ID":"9fb740d0-8b1d-4e74-980c-417f9b4cc855","Type":"ContainerDied","Data":"f34ca3254e5e46aee4bddda2fbb4b53cc03a000333b4add97aa0cdb7e64dcbf9"} Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.040976 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g8v8d" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="registry-server" probeResult="failure" output=< Feb 20 00:12:11 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Feb 20 00:12:11 crc kubenswrapper[4750]: > Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.183023 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.212326 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j"] Feb 20 00:12:11 crc kubenswrapper[4750]: E0220 00:12:11.212861 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="registry-server" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.212896 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="registry-server" Feb 20 00:12:11 crc kubenswrapper[4750]: E0220 00:12:11.212911 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb740d0-8b1d-4e74-980c-417f9b4cc855" containerName="route-controller-manager" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.212919 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb740d0-8b1d-4e74-980c-417f9b4cc855" containerName="route-controller-manager" Feb 20 00:12:11 crc kubenswrapper[4750]: E0220 00:12:11.212935 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="extract-content" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.212942 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="extract-content" Feb 20 00:12:11 crc kubenswrapper[4750]: E0220 00:12:11.212952 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="extract-utilities" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.212960 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="extract-utilities" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.228279 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd8558f-eb0c-4f43-aa1f-809ec986a7a7" containerName="registry-server" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.228330 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb740d0-8b1d-4e74-980c-417f9b4cc855" containerName="route-controller-manager" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.228906 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j"] Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.229082 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382312 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb740d0-8b1d-4e74-980c-417f9b4cc855-serving-cert\") pod \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382362 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-config\") pod \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382398 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-client-ca\") pod \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382425 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m498z\" (UniqueName: \"kubernetes.io/projected/9fb740d0-8b1d-4e74-980c-417f9b4cc855-kube-api-access-m498z\") pod \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\" (UID: \"9fb740d0-8b1d-4e74-980c-417f9b4cc855\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382651 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7450b6-6745-4758-babb-ef969b62e570-serving-cert\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382681 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-client-ca\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382705 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-config\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.382722 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf4mn\" (UniqueName: \"kubernetes.io/projected/8e7450b6-6745-4758-babb-ef969b62e570-kube-api-access-rf4mn\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.383100 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-config" (OuterVolumeSpecName: "config") pod "9fb740d0-8b1d-4e74-980c-417f9b4cc855" (UID: "9fb740d0-8b1d-4e74-980c-417f9b4cc855"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.383150 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-client-ca" (OuterVolumeSpecName: "client-ca") pod "9fb740d0-8b1d-4e74-980c-417f9b4cc855" (UID: "9fb740d0-8b1d-4e74-980c-417f9b4cc855"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.388290 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb740d0-8b1d-4e74-980c-417f9b4cc855-kube-api-access-m498z" (OuterVolumeSpecName: "kube-api-access-m498z") pod "9fb740d0-8b1d-4e74-980c-417f9b4cc855" (UID: "9fb740d0-8b1d-4e74-980c-417f9b4cc855"). InnerVolumeSpecName "kube-api-access-m498z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.390274 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb740d0-8b1d-4e74-980c-417f9b4cc855-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9fb740d0-8b1d-4e74-980c-417f9b4cc855" (UID: "9fb740d0-8b1d-4e74-980c-417f9b4cc855"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.416276 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.483960 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-client-ca\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484010 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7450b6-6745-4758-babb-ef969b62e570-serving-cert\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484044 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-config\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484069 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf4mn\" (UniqueName: \"kubernetes.io/projected/8e7450b6-6745-4758-babb-ef969b62e570-kube-api-access-rf4mn\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484163 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9fb740d0-8b1d-4e74-980c-417f9b4cc855-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484184 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484194 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9fb740d0-8b1d-4e74-980c-417f9b4cc855-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484204 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m498z\" (UniqueName: \"kubernetes.io/projected/9fb740d0-8b1d-4e74-980c-417f9b4cc855-kube-api-access-m498z\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.484990 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-client-ca\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.485486 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-config\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.498935 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7450b6-6745-4758-babb-ef969b62e570-serving-cert\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.502022 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf4mn\" (UniqueName: \"kubernetes.io/projected/8e7450b6-6745-4758-babb-ef969b62e570-kube-api-access-rf4mn\") pod \"route-controller-manager-64d4f9dcdd-qpp7j\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.553231 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.584789 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-config\") pod \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.584877 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-client-ca\") pod \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.584925 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pgv5\" (UniqueName: \"kubernetes.io/projected/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-kube-api-access-7pgv5\") pod \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.584965 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-serving-cert\") pod \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.584996 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-proxy-ca-bundles\") pod \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\" (UID: \"92e9fbea-27b0-4b5f-a4d8-6a014836dad1\") " Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.585811 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-client-ca" (OuterVolumeSpecName: "client-ca") pod "92e9fbea-27b0-4b5f-a4d8-6a014836dad1" (UID: "92e9fbea-27b0-4b5f-a4d8-6a014836dad1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.585823 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "92e9fbea-27b0-4b5f-a4d8-6a014836dad1" (UID: "92e9fbea-27b0-4b5f-a4d8-6a014836dad1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.585870 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-config" (OuterVolumeSpecName: "config") pod "92e9fbea-27b0-4b5f-a4d8-6a014836dad1" (UID: "92e9fbea-27b0-4b5f-a4d8-6a014836dad1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.588273 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-kube-api-access-7pgv5" (OuterVolumeSpecName: "kube-api-access-7pgv5") pod "92e9fbea-27b0-4b5f-a4d8-6a014836dad1" (UID: "92e9fbea-27b0-4b5f-a4d8-6a014836dad1"). InnerVolumeSpecName "kube-api-access-7pgv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.598403 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "92e9fbea-27b0-4b5f-a4d8-6a014836dad1" (UID: "92e9fbea-27b0-4b5f-a4d8-6a014836dad1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.686462 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.686492 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.686508 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pgv5\" (UniqueName: \"kubernetes.io/projected/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-kube-api-access-7pgv5\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.686525 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.686534 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/92e9fbea-27b0-4b5f-a4d8-6a014836dad1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.807597 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j"] Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.873525 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" event={"ID":"9fb740d0-8b1d-4e74-980c-417f9b4cc855","Type":"ContainerDied","Data":"0e41b0378b95b13c12080cb41cffa68a71f5ac90f216e8719f3a7da8b2ec4900"} Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.873542 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.873610 4750 scope.go:117] "RemoveContainer" containerID="f34ca3254e5e46aee4bddda2fbb4b53cc03a000333b4add97aa0cdb7e64dcbf9" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.874571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" event={"ID":"8e7450b6-6745-4758-babb-ef969b62e570","Type":"ContainerStarted","Data":"26a70e40cb8b8b39783efb087faa4b3bde77462b75594268c5932ab4d3ec0bbe"} Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.894798 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" event={"ID":"92e9fbea-27b0-4b5f-a4d8-6a014836dad1","Type":"ContainerDied","Data":"e18e65dd30f66a1c82300289a30c6df893fdcd06342ce806bcfd5868f94566b7"} Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.894896 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79df6c4bb4-vtb65" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.908773 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv"] Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.910954 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5b67bccc-f26qv"] Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.930659 4750 scope.go:117] "RemoveContainer" containerID="b2b3629b491bea280dcc67153335ea208031b2451293cbb046bc0127c2005dff" Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.930786 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79df6c4bb4-vtb65"] Feb 20 00:12:11 crc kubenswrapper[4750]: I0220 00:12:11.932640 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-79df6c4bb4-vtb65"] Feb 20 00:12:12 crc kubenswrapper[4750]: I0220 00:12:12.566164 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92e9fbea-27b0-4b5f-a4d8-6a014836dad1" path="/var/lib/kubelet/pods/92e9fbea-27b0-4b5f-a4d8-6a014836dad1/volumes" Feb 20 00:12:12 crc kubenswrapper[4750]: I0220 00:12:12.567440 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb740d0-8b1d-4e74-980c-417f9b4cc855" path="/var/lib/kubelet/pods/9fb740d0-8b1d-4e74-980c-417f9b4cc855/volumes" Feb 20 00:12:12 crc kubenswrapper[4750]: I0220 00:12:12.901557 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" event={"ID":"8e7450b6-6745-4758-babb-ef969b62e570","Type":"ContainerStarted","Data":"999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d"} Feb 20 00:12:12 crc kubenswrapper[4750]: I0220 00:12:12.901934 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:12 crc kubenswrapper[4750]: I0220 00:12:12.909457 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:12 crc kubenswrapper[4750]: I0220 00:12:12.922512 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" podStartSLOduration=2.92249434 podStartE2EDuration="2.92249434s" podCreationTimestamp="2026-02-20 00:12:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:12:12.922019747 +0000 UTC m=+217.116855796" watchObservedRunningTime="2026-02-20 00:12:12.92249434 +0000 UTC m=+217.117330389" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.739017 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6fb6cf77dd-gntht"] Feb 20 00:12:13 crc kubenswrapper[4750]: E0220 00:12:13.739316 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e9fbea-27b0-4b5f-a4d8-6a014836dad1" containerName="controller-manager" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.739338 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e9fbea-27b0-4b5f-a4d8-6a014836dad1" containerName="controller-manager" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.739452 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e9fbea-27b0-4b5f-a4d8-6a014836dad1" containerName="controller-manager" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.739908 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.742891 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.743161 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.743546 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.745017 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.746329 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.750421 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fb6cf77dd-gntht"] Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.754358 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.760835 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.813788 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-client-ca\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.813839 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-config\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.813869 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkhzn\" (UniqueName: \"kubernetes.io/projected/af9de212-a200-40f4-9a9f-c9603c74099c-kube-api-access-gkhzn\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.813896 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-proxy-ca-bundles\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.813935 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9de212-a200-40f4-9a9f-c9603c74099c-serving-cert\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.915159 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-client-ca\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.915205 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-config\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.915229 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkhzn\" (UniqueName: \"kubernetes.io/projected/af9de212-a200-40f4-9a9f-c9603c74099c-kube-api-access-gkhzn\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.915253 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-proxy-ca-bundles\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.915287 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9de212-a200-40f4-9a9f-c9603c74099c-serving-cert\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.916517 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-client-ca\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.916914 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-config\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.918975 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-proxy-ca-bundles\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.922143 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9de212-a200-40f4-9a9f-c9603c74099c-serving-cert\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:13 crc kubenswrapper[4750]: I0220 00:12:13.942084 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkhzn\" (UniqueName: \"kubernetes.io/projected/af9de212-a200-40f4-9a9f-c9603c74099c-kube-api-access-gkhzn\") pod \"controller-manager-6fb6cf77dd-gntht\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:14 crc kubenswrapper[4750]: I0220 00:12:14.056774 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:14 crc kubenswrapper[4750]: I0220 00:12:14.290973 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6fb6cf77dd-gntht"] Feb 20 00:12:14 crc kubenswrapper[4750]: I0220 00:12:14.915845 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" event={"ID":"af9de212-a200-40f4-9a9f-c9603c74099c","Type":"ContainerStarted","Data":"60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c"} Feb 20 00:12:14 crc kubenswrapper[4750]: I0220 00:12:14.916219 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" event={"ID":"af9de212-a200-40f4-9a9f-c9603c74099c","Type":"ContainerStarted","Data":"a8b62f82f694d5d314a0240d5e955aa159e4fc848237645af84c59d2b7253cf2"} Feb 20 00:12:14 crc kubenswrapper[4750]: I0220 00:12:14.944478 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" podStartSLOduration=4.944449585 podStartE2EDuration="4.944449585s" podCreationTimestamp="2026-02-20 00:12:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:12:14.939378926 +0000 UTC m=+219.134214995" watchObservedRunningTime="2026-02-20 00:12:14.944449585 +0000 UTC m=+219.139285644" Feb 20 00:12:15 crc kubenswrapper[4750]: I0220 00:12:15.921155 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:15 crc kubenswrapper[4750]: I0220 00:12:15.926080 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:16 crc kubenswrapper[4750]: I0220 00:12:16.408697 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:12:16 crc kubenswrapper[4750]: I0220 00:12:16.545426 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:12:16 crc kubenswrapper[4750]: I0220 00:12:16.545466 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:12:16 crc kubenswrapper[4750]: I0220 00:12:16.590729 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:12:16 crc kubenswrapper[4750]: I0220 00:12:16.966969 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:12:16 crc kubenswrapper[4750]: I0220 00:12:16.967267 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:12:16 crc kubenswrapper[4750]: I0220 00:12:16.979465 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:12:17 crc kubenswrapper[4750]: I0220 00:12:17.040651 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:12:17 crc kubenswrapper[4750]: I0220 00:12:17.986987 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:12:18 crc kubenswrapper[4750]: I0220 00:12:18.630591 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5cq9b"] Feb 20 00:12:19 crc kubenswrapper[4750]: I0220 00:12:19.568415 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:12:19 crc kubenswrapper[4750]: I0220 00:12:19.568682 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:12:19 crc kubenswrapper[4750]: I0220 00:12:19.634348 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:12:19 crc kubenswrapper[4750]: I0220 00:12:19.948287 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5cq9b" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="registry-server" containerID="cri-o://6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5" gracePeriod=2 Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.025162 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.025596 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.132796 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.436766 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.634697 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sjhf\" (UniqueName: \"kubernetes.io/projected/ac3d2b05-2e15-4058-a6f1-1ef965755b02-kube-api-access-2sjhf\") pod \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.634845 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-utilities\") pod \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.635266 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-catalog-content\") pod \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\" (UID: \"ac3d2b05-2e15-4058-a6f1-1ef965755b02\") " Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.635638 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-utilities" (OuterVolumeSpecName: "utilities") pod "ac3d2b05-2e15-4058-a6f1-1ef965755b02" (UID: "ac3d2b05-2e15-4058-a6f1-1ef965755b02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.635850 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.639492 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac3d2b05-2e15-4058-a6f1-1ef965755b02-kube-api-access-2sjhf" (OuterVolumeSpecName: "kube-api-access-2sjhf") pod "ac3d2b05-2e15-4058-a6f1-1ef965755b02" (UID: "ac3d2b05-2e15-4058-a6f1-1ef965755b02"). InnerVolumeSpecName "kube-api-access-2sjhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.691408 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac3d2b05-2e15-4058-a6f1-1ef965755b02" (UID: "ac3d2b05-2e15-4058-a6f1-1ef965755b02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.736670 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac3d2b05-2e15-4058-a6f1-1ef965755b02-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.736707 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sjhf\" (UniqueName: \"kubernetes.io/projected/ac3d2b05-2e15-4058-a6f1-1ef965755b02-kube-api-access-2sjhf\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.954850 4750 generic.go:334] "Generic (PLEG): container finished" podID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerID="6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5" exitCode=0 Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.954942 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5cq9b" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.954998 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5cq9b" event={"ID":"ac3d2b05-2e15-4058-a6f1-1ef965755b02","Type":"ContainerDied","Data":"6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5"} Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.955050 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5cq9b" event={"ID":"ac3d2b05-2e15-4058-a6f1-1ef965755b02","Type":"ContainerDied","Data":"de3feaccd4a83250e0b976aafd730794df8be48006f3badfa156163dd9b65c23"} Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.955075 4750 scope.go:117] "RemoveContainer" containerID="6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.975711 4750 scope.go:117] "RemoveContainer" containerID="921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886" Feb 20 00:12:20 crc kubenswrapper[4750]: I0220 00:12:20.999740 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5cq9b"] Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.002093 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5cq9b"] Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.011871 4750 scope.go:117] "RemoveContainer" containerID="b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5" Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.031410 4750 scope.go:117] "RemoveContainer" containerID="6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5" Feb 20 00:12:21 crc kubenswrapper[4750]: E0220 00:12:21.031964 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5\": container with ID starting with 6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5 not found: ID does not exist" containerID="6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5" Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.032017 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5"} err="failed to get container status \"6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5\": rpc error: code = NotFound desc = could not find container \"6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5\": container with ID starting with 6d9a881211e52f1850589cfd0d0a3a23c3ced713246510514d5cc49284ffc6b5 not found: ID does not exist" Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.032050 4750 scope.go:117] "RemoveContainer" containerID="921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886" Feb 20 00:12:21 crc kubenswrapper[4750]: E0220 00:12:21.032473 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886\": container with ID starting with 921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886 not found: ID does not exist" containerID="921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886" Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.032521 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886"} err="failed to get container status \"921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886\": rpc error: code = NotFound desc = could not find container \"921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886\": container with ID starting with 921bc67792877e04aa69ae1ea19dc6413c724ea28825541704862d22a9e74886 not found: ID does not exist" Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.032551 4750 scope.go:117] "RemoveContainer" containerID="b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5" Feb 20 00:12:21 crc kubenswrapper[4750]: E0220 00:12:21.033161 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5\": container with ID starting with b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5 not found: ID does not exist" containerID="b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5" Feb 20 00:12:21 crc kubenswrapper[4750]: I0220 00:12:21.033260 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5"} err="failed to get container status \"b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5\": rpc error: code = NotFound desc = could not find container \"b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5\": container with ID starting with b0e36717fd25855fa01d7f0606b90ffbb7a49a3995f6bb4593eb26f9b57c43d5 not found: ID does not exist" Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.571052 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" path="/var/lib/kubelet/pods/ac3d2b05-2e15-4058-a6f1-1ef965755b02/volumes" Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.830678 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8v8d"] Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.831014 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8v8d" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="registry-server" containerID="cri-o://7706116e03096c00aeeb051fcc8954f37e220c90098d9478e4ee5b3ef814f88a" gracePeriod=2 Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.832976 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" podUID="c4c40a07-5939-4d83-83af-e22f773ee511" containerName="oauth-openshift" containerID="cri-o://68f0334058ba4fd26024c15c83c64f230975ff080336f93bb46820f3921ea075" gracePeriod=15 Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.974952 4750 generic.go:334] "Generic (PLEG): container finished" podID="c4c40a07-5939-4d83-83af-e22f773ee511" containerID="68f0334058ba4fd26024c15c83c64f230975ff080336f93bb46820f3921ea075" exitCode=0 Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.975036 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" event={"ID":"c4c40a07-5939-4d83-83af-e22f773ee511","Type":"ContainerDied","Data":"68f0334058ba4fd26024c15c83c64f230975ff080336f93bb46820f3921ea075"} Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.981612 4750 generic.go:334] "Generic (PLEG): container finished" podID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerID="7706116e03096c00aeeb051fcc8954f37e220c90098d9478e4ee5b3ef814f88a" exitCode=0 Feb 20 00:12:22 crc kubenswrapper[4750]: I0220 00:12:22.981659 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8v8d" event={"ID":"e21f8aa9-f543-4f68-b241-fde6f0e6bb49","Type":"ContainerDied","Data":"7706116e03096c00aeeb051fcc8954f37e220c90098d9478e4ee5b3ef814f88a"} Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.457902 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.463308 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474135 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-utilities\") pod \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474206 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-catalog-content\") pod \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474239 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-cliconfig\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474264 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-login\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474318 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w5wt\" (UniqueName: \"kubernetes.io/projected/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-kube-api-access-5w5wt\") pod \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\" (UID: \"e21f8aa9-f543-4f68-b241-fde6f0e6bb49\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474350 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-audit-policies\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474385 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-error\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474409 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-session\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474429 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4c40a07-5939-4d83-83af-e22f773ee511-audit-dir\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474458 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-provider-selection\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474501 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-idp-0-file-data\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474526 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-ocp-branding-template\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474559 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-serving-cert\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474588 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwnhs\" (UniqueName: \"kubernetes.io/projected/c4c40a07-5939-4d83-83af-e22f773ee511-kube-api-access-lwnhs\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474610 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-router-certs\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474635 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-service-ca\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.474658 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-trusted-ca-bundle\") pod \"c4c40a07-5939-4d83-83af-e22f773ee511\" (UID: \"c4c40a07-5939-4d83-83af-e22f773ee511\") " Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.475196 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.475416 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-utilities" (OuterVolumeSpecName: "utilities") pod "e21f8aa9-f543-4f68-b241-fde6f0e6bb49" (UID: "e21f8aa9-f543-4f68-b241-fde6f0e6bb49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.475458 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.477083 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c4c40a07-5939-4d83-83af-e22f773ee511-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.477579 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.477650 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.481497 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.493951 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.494456 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.494623 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.494898 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.497945 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.498246 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.498323 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4c40a07-5939-4d83-83af-e22f773ee511-kube-api-access-lwnhs" (OuterVolumeSpecName: "kube-api-access-lwnhs") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "kube-api-access-lwnhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.498385 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-kube-api-access-5w5wt" (OuterVolumeSpecName: "kube-api-access-5w5wt") pod "e21f8aa9-f543-4f68-b241-fde6f0e6bb49" (UID: "e21f8aa9-f543-4f68-b241-fde6f0e6bb49"). InnerVolumeSpecName "kube-api-access-5w5wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.531541 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c4c40a07-5939-4d83-83af-e22f773ee511" (UID: "c4c40a07-5939-4d83-83af-e22f773ee511"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575349 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575377 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwnhs\" (UniqueName: \"kubernetes.io/projected/c4c40a07-5939-4d83-83af-e22f773ee511-kube-api-access-lwnhs\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575387 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575397 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575407 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575418 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575428 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575438 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575446 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w5wt\" (UniqueName: \"kubernetes.io/projected/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-kube-api-access-5w5wt\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575455 4750 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4c40a07-5939-4d83-83af-e22f773ee511-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575463 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575470 4750 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4c40a07-5939-4d83-83af-e22f773ee511-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575486 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575494 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575503 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.575511 4750 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4c40a07-5939-4d83-83af-e22f773ee511-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.668471 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e21f8aa9-f543-4f68-b241-fde6f0e6bb49" (UID: "e21f8aa9-f543-4f68-b241-fde6f0e6bb49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.676785 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21f8aa9-f543-4f68-b241-fde6f0e6bb49-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.989544 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" event={"ID":"c4c40a07-5939-4d83-83af-e22f773ee511","Type":"ContainerDied","Data":"d2a8eddfa669d16426be47cc1c342eb9b84d6a80fcca13fc3d3aca9f93b8f29b"} Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.989597 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nd2j6" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.989623 4750 scope.go:117] "RemoveContainer" containerID="68f0334058ba4fd26024c15c83c64f230975ff080336f93bb46820f3921ea075" Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.992880 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8v8d" event={"ID":"e21f8aa9-f543-4f68-b241-fde6f0e6bb49","Type":"ContainerDied","Data":"2a389c259fadf83772f601d5a047fec956873874f61d4307ca3f487639f4d68f"} Feb 20 00:12:23 crc kubenswrapper[4750]: I0220 00:12:23.992994 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8v8d" Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.014427 4750 scope.go:117] "RemoveContainer" containerID="7706116e03096c00aeeb051fcc8954f37e220c90098d9478e4ee5b3ef814f88a" Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.042399 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nd2j6"] Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.046532 4750 scope.go:117] "RemoveContainer" containerID="78a18902442e10656e64a818e69f07e80541f35e124c18336c58b9dd36e1dec1" Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.048444 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nd2j6"] Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.066157 4750 scope.go:117] "RemoveContainer" containerID="f3c4eee74f8e7c7e8d8b74e47b77b3a4c38749577ddf297e570811825a96e30f" Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.069745 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8v8d"] Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.072414 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8v8d"] Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.570909 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4c40a07-5939-4d83-83af-e22f773ee511" path="/var/lib/kubelet/pods/c4c40a07-5939-4d83-83af-e22f773ee511/volumes" Feb 20 00:12:24 crc kubenswrapper[4750]: I0220 00:12:24.571917 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" path="/var/lib/kubelet/pods/e21f8aa9-f543-4f68-b241-fde6f0e6bb49/volumes" Feb 20 00:12:26 crc kubenswrapper[4750]: I0220 00:12:26.670235 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:12:26 crc kubenswrapper[4750]: I0220 00:12:26.670569 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:12:26 crc kubenswrapper[4750]: I0220 00:12:26.670631 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:12:26 crc kubenswrapper[4750]: I0220 00:12:26.671409 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952"} pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 00:12:26 crc kubenswrapper[4750]: I0220 00:12:26.671512 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" containerID="cri-o://67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952" gracePeriod=600 Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.019157 4750 generic.go:334] "Generic (PLEG): container finished" podID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerID="67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952" exitCode=0 Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.019551 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerDied","Data":"67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952"} Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.019584 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"7f3a4507cbe00af55a5fb388980bbf31c52a319e48434ce9fcf1faccc79339c1"} Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749368 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5b8f568f87-66885"] Feb 20 00:12:27 crc kubenswrapper[4750]: E0220 00:12:27.749569 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="registry-server" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749580 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="registry-server" Feb 20 00:12:27 crc kubenswrapper[4750]: E0220 00:12:27.749591 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="extract-content" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749598 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="extract-content" Feb 20 00:12:27 crc kubenswrapper[4750]: E0220 00:12:27.749610 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="extract-utilities" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749617 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="extract-utilities" Feb 20 00:12:27 crc kubenswrapper[4750]: E0220 00:12:27.749624 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="extract-utilities" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749629 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="extract-utilities" Feb 20 00:12:27 crc kubenswrapper[4750]: E0220 00:12:27.749638 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c40a07-5939-4d83-83af-e22f773ee511" containerName="oauth-openshift" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749644 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c40a07-5939-4d83-83af-e22f773ee511" containerName="oauth-openshift" Feb 20 00:12:27 crc kubenswrapper[4750]: E0220 00:12:27.749652 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="extract-content" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749658 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="extract-content" Feb 20 00:12:27 crc kubenswrapper[4750]: E0220 00:12:27.749666 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="registry-server" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749672 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="registry-server" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749762 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4c40a07-5939-4d83-83af-e22f773ee511" containerName="oauth-openshift" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749772 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="e21f8aa9-f543-4f68-b241-fde6f0e6bb49" containerName="registry-server" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.749785 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac3d2b05-2e15-4058-a6f1-1ef965755b02" containerName="registry-server" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.750228 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.757183 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.757352 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.758072 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.758880 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.759042 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.759154 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.759275 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.759394 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.759661 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.759753 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.759854 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.763876 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.770441 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.773431 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5b8f568f87-66885"] Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.773819 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.780331 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.832885 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/434fbf8f-103c-43f3-8b7c-dcbf8a051166-audit-dir\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.832972 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833023 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833075 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-error\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833111 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-service-ca\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833170 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzp56\" (UniqueName: \"kubernetes.io/projected/434fbf8f-103c-43f3-8b7c-dcbf8a051166-kube-api-access-vzp56\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833218 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833254 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-login\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833292 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833325 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-router-certs\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833367 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833411 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-audit-policies\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833480 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.833527 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-session\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.934712 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-session\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.934781 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/434fbf8f-103c-43f3-8b7c-dcbf8a051166-audit-dir\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.934823 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.934855 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.934902 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-error\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.934908 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/434fbf8f-103c-43f3-8b7c-dcbf8a051166-audit-dir\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.934928 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzp56\" (UniqueName: \"kubernetes.io/projected/434fbf8f-103c-43f3-8b7c-dcbf8a051166-kube-api-access-vzp56\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935005 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-service-ca\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935058 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935100 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-login\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935177 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935250 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-router-certs\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935293 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935342 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-audit-policies\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.935380 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.937386 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.937518 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-service-ca\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.938332 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.939070 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/434fbf8f-103c-43f3-8b7c-dcbf8a051166-audit-policies\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.942647 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.943469 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-login\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.943563 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.943996 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.944347 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-session\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.944372 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-router-certs\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.944787 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-user-template-error\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:27 crc kubenswrapper[4750]: I0220 00:12:27.944827 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/434fbf8f-103c-43f3-8b7c-dcbf8a051166-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:28 crc kubenswrapper[4750]: I0220 00:12:28.024017 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzp56\" (UniqueName: \"kubernetes.io/projected/434fbf8f-103c-43f3-8b7c-dcbf8a051166-kube-api-access-vzp56\") pod \"oauth-openshift-5b8f568f87-66885\" (UID: \"434fbf8f-103c-43f3-8b7c-dcbf8a051166\") " pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:28 crc kubenswrapper[4750]: I0220 00:12:28.064544 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:28 crc kubenswrapper[4750]: I0220 00:12:28.486863 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5b8f568f87-66885"] Feb 20 00:12:28 crc kubenswrapper[4750]: W0220 00:12:28.493983 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod434fbf8f_103c_43f3_8b7c_dcbf8a051166.slice/crio-76090f82de1ecd39c667c8e8b1b9dfb02caca582804ceec531d15025747e4841 WatchSource:0}: Error finding container 76090f82de1ecd39c667c8e8b1b9dfb02caca582804ceec531d15025747e4841: Status 404 returned error can't find the container with id 76090f82de1ecd39c667c8e8b1b9dfb02caca582804ceec531d15025747e4841 Feb 20 00:12:29 crc kubenswrapper[4750]: I0220 00:12:29.041893 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" event={"ID":"434fbf8f-103c-43f3-8b7c-dcbf8a051166","Type":"ContainerStarted","Data":"4731dacff3077d185ed215e0842aaea6f048accc4aa2dfa089a92d4bdcfabab4"} Feb 20 00:12:29 crc kubenswrapper[4750]: I0220 00:12:29.042315 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" event={"ID":"434fbf8f-103c-43f3-8b7c-dcbf8a051166","Type":"ContainerStarted","Data":"76090f82de1ecd39c667c8e8b1b9dfb02caca582804ceec531d15025747e4841"} Feb 20 00:12:29 crc kubenswrapper[4750]: I0220 00:12:29.042345 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:29 crc kubenswrapper[4750]: I0220 00:12:29.068486 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" podStartSLOduration=32.068461198 podStartE2EDuration="32.068461198s" podCreationTimestamp="2026-02-20 00:11:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:12:29.067390298 +0000 UTC m=+233.262226397" watchObservedRunningTime="2026-02-20 00:12:29.068461198 +0000 UTC m=+233.263297247" Feb 20 00:12:29 crc kubenswrapper[4750]: I0220 00:12:29.110002 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5b8f568f87-66885" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.014769 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6fb6cf77dd-gntht"] Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.015347 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" podUID="af9de212-a200-40f4-9a9f-c9603c74099c" containerName="controller-manager" containerID="cri-o://60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c" gracePeriod=30 Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.104470 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j"] Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.104669 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" podUID="8e7450b6-6745-4758-babb-ef969b62e570" containerName="route-controller-manager" containerID="cri-o://999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d" gracePeriod=30 Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.617925 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.621917 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676664 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-client-ca\") pod \"8e7450b6-6745-4758-babb-ef969b62e570\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676747 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf4mn\" (UniqueName: \"kubernetes.io/projected/8e7450b6-6745-4758-babb-ef969b62e570-kube-api-access-rf4mn\") pod \"8e7450b6-6745-4758-babb-ef969b62e570\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676774 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9de212-a200-40f4-9a9f-c9603c74099c-serving-cert\") pod \"af9de212-a200-40f4-9a9f-c9603c74099c\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676804 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7450b6-6745-4758-babb-ef969b62e570-serving-cert\") pod \"8e7450b6-6745-4758-babb-ef969b62e570\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676845 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-config\") pod \"8e7450b6-6745-4758-babb-ef969b62e570\" (UID: \"8e7450b6-6745-4758-babb-ef969b62e570\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676878 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-proxy-ca-bundles\") pod \"af9de212-a200-40f4-9a9f-c9603c74099c\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676932 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-config\") pod \"af9de212-a200-40f4-9a9f-c9603c74099c\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.676964 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkhzn\" (UniqueName: \"kubernetes.io/projected/af9de212-a200-40f4-9a9f-c9603c74099c-kube-api-access-gkhzn\") pod \"af9de212-a200-40f4-9a9f-c9603c74099c\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.677002 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-client-ca\") pod \"af9de212-a200-40f4-9a9f-c9603c74099c\" (UID: \"af9de212-a200-40f4-9a9f-c9603c74099c\") " Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678194 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-client-ca" (OuterVolumeSpecName: "client-ca") pod "af9de212-a200-40f4-9a9f-c9603c74099c" (UID: "af9de212-a200-40f4-9a9f-c9603c74099c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678209 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "af9de212-a200-40f4-9a9f-c9603c74099c" (UID: "af9de212-a200-40f4-9a9f-c9603c74099c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678216 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-config" (OuterVolumeSpecName: "config") pod "8e7450b6-6745-4758-babb-ef969b62e570" (UID: "8e7450b6-6745-4758-babb-ef969b62e570"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678345 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-client-ca" (OuterVolumeSpecName: "client-ca") pod "8e7450b6-6745-4758-babb-ef969b62e570" (UID: "8e7450b6-6745-4758-babb-ef969b62e570"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678759 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678781 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e7450b6-6745-4758-babb-ef969b62e570-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678790 4750 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678801 4750 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.678864 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-config" (OuterVolumeSpecName: "config") pod "af9de212-a200-40f4-9a9f-c9603c74099c" (UID: "af9de212-a200-40f4-9a9f-c9603c74099c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.683523 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e7450b6-6745-4758-babb-ef969b62e570-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8e7450b6-6745-4758-babb-ef969b62e570" (UID: "8e7450b6-6745-4758-babb-ef969b62e570"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.683532 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af9de212-a200-40f4-9a9f-c9603c74099c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "af9de212-a200-40f4-9a9f-c9603c74099c" (UID: "af9de212-a200-40f4-9a9f-c9603c74099c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.683591 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9de212-a200-40f4-9a9f-c9603c74099c-kube-api-access-gkhzn" (OuterVolumeSpecName: "kube-api-access-gkhzn") pod "af9de212-a200-40f4-9a9f-c9603c74099c" (UID: "af9de212-a200-40f4-9a9f-c9603c74099c"). InnerVolumeSpecName "kube-api-access-gkhzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.685240 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e7450b6-6745-4758-babb-ef969b62e570-kube-api-access-rf4mn" (OuterVolumeSpecName: "kube-api-access-rf4mn") pod "8e7450b6-6745-4758-babb-ef969b62e570" (UID: "8e7450b6-6745-4758-babb-ef969b62e570"). InnerVolumeSpecName "kube-api-access-rf4mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.779632 4750 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af9de212-a200-40f4-9a9f-c9603c74099c-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.779659 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkhzn\" (UniqueName: \"kubernetes.io/projected/af9de212-a200-40f4-9a9f-c9603c74099c-kube-api-access-gkhzn\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.779669 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf4mn\" (UniqueName: \"kubernetes.io/projected/8e7450b6-6745-4758-babb-ef969b62e570-kube-api-access-rf4mn\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.779678 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af9de212-a200-40f4-9a9f-c9603c74099c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:30 crc kubenswrapper[4750]: I0220 00:12:30.779688 4750 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e7450b6-6745-4758-babb-ef969b62e570-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.053549 4750 generic.go:334] "Generic (PLEG): container finished" podID="af9de212-a200-40f4-9a9f-c9603c74099c" containerID="60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c" exitCode=0 Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.053650 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" event={"ID":"af9de212-a200-40f4-9a9f-c9603c74099c","Type":"ContainerDied","Data":"60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c"} Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.053690 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" event={"ID":"af9de212-a200-40f4-9a9f-c9603c74099c","Type":"ContainerDied","Data":"a8b62f82f694d5d314a0240d5e955aa159e4fc848237645af84c59d2b7253cf2"} Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.053720 4750 scope.go:117] "RemoveContainer" containerID="60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.053857 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6fb6cf77dd-gntht" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.057920 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.058027 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" event={"ID":"8e7450b6-6745-4758-babb-ef969b62e570","Type":"ContainerDied","Data":"999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d"} Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.057808 4750 generic.go:334] "Generic (PLEG): container finished" podID="8e7450b6-6745-4758-babb-ef969b62e570" containerID="999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d" exitCode=0 Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.059734 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j" event={"ID":"8e7450b6-6745-4758-babb-ef969b62e570","Type":"ContainerDied","Data":"26a70e40cb8b8b39783efb087faa4b3bde77462b75594268c5932ab4d3ec0bbe"} Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.091699 4750 scope.go:117] "RemoveContainer" containerID="60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c" Feb 20 00:12:31 crc kubenswrapper[4750]: E0220 00:12:31.092802 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c\": container with ID starting with 60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c not found: ID does not exist" containerID="60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.092838 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c"} err="failed to get container status \"60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c\": rpc error: code = NotFound desc = could not find container \"60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c\": container with ID starting with 60f26bdb3c90857d941f272f194f4c422d0ac38a40267f99703af3f8165ad08c not found: ID does not exist" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.092864 4750 scope.go:117] "RemoveContainer" containerID="999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.105228 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j"] Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.113787 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64d4f9dcdd-qpp7j"] Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.118249 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6fb6cf77dd-gntht"] Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.120593 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6fb6cf77dd-gntht"] Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.131872 4750 scope.go:117] "RemoveContainer" containerID="999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d" Feb 20 00:12:31 crc kubenswrapper[4750]: E0220 00:12:31.132628 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d\": container with ID starting with 999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d not found: ID does not exist" containerID="999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.132683 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d"} err="failed to get container status \"999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d\": rpc error: code = NotFound desc = could not find container \"999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d\": container with ID starting with 999dc4ca9a52e3cd51acb092be18118e9838bf9a850825ed0614c720b092ad5d not found: ID does not exist" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.758154 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr"] Feb 20 00:12:31 crc kubenswrapper[4750]: E0220 00:12:31.758899 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9de212-a200-40f4-9a9f-c9603c74099c" containerName="controller-manager" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.758929 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9de212-a200-40f4-9a9f-c9603c74099c" containerName="controller-manager" Feb 20 00:12:31 crc kubenswrapper[4750]: E0220 00:12:31.758997 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7450b6-6745-4758-babb-ef969b62e570" containerName="route-controller-manager" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.759016 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7450b6-6745-4758-babb-ef969b62e570" containerName="route-controller-manager" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.759729 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9de212-a200-40f4-9a9f-c9603c74099c" containerName="controller-manager" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.759779 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e7450b6-6745-4758-babb-ef969b62e570" containerName="route-controller-manager" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.760633 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d5d6d49db-pzb99"] Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.763440 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.763685 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.771016 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.771564 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.772070 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.772796 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.772976 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.773324 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.773436 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.773610 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.773768 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.773801 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.773963 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.774018 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.784254 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.788273 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr"] Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795185 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28914972-d1a5-44a1-914a-326579886426-serving-cert\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795251 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-client-ca\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795340 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28914972-d1a5-44a1-914a-326579886426-config\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795384 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-config\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795424 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-proxy-ca-bundles\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795488 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28914972-d1a5-44a1-914a-326579886426-client-ca\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795527 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw8g9\" (UniqueName: \"kubernetes.io/projected/0d112128-f54d-4045-88bc-3ddac499baab-kube-api-access-mw8g9\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795603 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qxp4\" (UniqueName: \"kubernetes.io/projected/28914972-d1a5-44a1-914a-326579886426-kube-api-access-5qxp4\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.795641 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d112128-f54d-4045-88bc-3ddac499baab-serving-cert\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.824524 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d5d6d49db-pzb99"] Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.896958 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qxp4\" (UniqueName: \"kubernetes.io/projected/28914972-d1a5-44a1-914a-326579886426-kube-api-access-5qxp4\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.897007 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d112128-f54d-4045-88bc-3ddac499baab-serving-cert\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.897076 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28914972-d1a5-44a1-914a-326579886426-serving-cert\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.897099 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-client-ca\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.897144 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28914972-d1a5-44a1-914a-326579886426-config\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.897162 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-config\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.897178 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-proxy-ca-bundles\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.898071 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28914972-d1a5-44a1-914a-326579886426-client-ca\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.898102 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw8g9\" (UniqueName: \"kubernetes.io/projected/0d112128-f54d-4045-88bc-3ddac499baab-kube-api-access-mw8g9\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.898017 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-client-ca\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.898956 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/28914972-d1a5-44a1-914a-326579886426-client-ca\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.899162 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28914972-d1a5-44a1-914a-326579886426-config\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.900210 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-proxy-ca-bundles\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.901011 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d112128-f54d-4045-88bc-3ddac499baab-config\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.904743 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28914972-d1a5-44a1-914a-326579886426-serving-cert\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.915249 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d112128-f54d-4045-88bc-3ddac499baab-serving-cert\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.916392 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw8g9\" (UniqueName: \"kubernetes.io/projected/0d112128-f54d-4045-88bc-3ddac499baab-kube-api-access-mw8g9\") pod \"controller-manager-7d5d6d49db-pzb99\" (UID: \"0d112128-f54d-4045-88bc-3ddac499baab\") " pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:31 crc kubenswrapper[4750]: I0220 00:12:31.923685 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qxp4\" (UniqueName: \"kubernetes.io/projected/28914972-d1a5-44a1-914a-326579886426-kube-api-access-5qxp4\") pod \"route-controller-manager-86f5897b9d-75pfr\" (UID: \"28914972-d1a5-44a1-914a-326579886426\") " pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:32 crc kubenswrapper[4750]: I0220 00:12:32.093653 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:32 crc kubenswrapper[4750]: I0220 00:12:32.101995 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:32 crc kubenswrapper[4750]: I0220 00:12:32.573436 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e7450b6-6745-4758-babb-ef969b62e570" path="/var/lib/kubelet/pods/8e7450b6-6745-4758-babb-ef969b62e570/volumes" Feb 20 00:12:32 crc kubenswrapper[4750]: I0220 00:12:32.574982 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9de212-a200-40f4-9a9f-c9603c74099c" path="/var/lib/kubelet/pods/af9de212-a200-40f4-9a9f-c9603c74099c/volumes" Feb 20 00:12:32 crc kubenswrapper[4750]: I0220 00:12:32.575845 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d5d6d49db-pzb99"] Feb 20 00:12:32 crc kubenswrapper[4750]: I0220 00:12:32.657322 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr"] Feb 20 00:12:32 crc kubenswrapper[4750]: W0220 00:12:32.668355 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28914972_d1a5_44a1_914a_326579886426.slice/crio-51789bf5af48aa6766a2de4863a3e89316bc453144205aba4bdb690f1923b2ef WatchSource:0}: Error finding container 51789bf5af48aa6766a2de4863a3e89316bc453144205aba4bdb690f1923b2ef: Status 404 returned error can't find the container with id 51789bf5af48aa6766a2de4863a3e89316bc453144205aba4bdb690f1923b2ef Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.077992 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" event={"ID":"28914972-d1a5-44a1-914a-326579886426","Type":"ContainerStarted","Data":"07b961fdc9c1341560124485b2f687fd55bae99ad32ea533c139d54b4a83b9a0"} Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.078063 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" event={"ID":"28914972-d1a5-44a1-914a-326579886426","Type":"ContainerStarted","Data":"51789bf5af48aa6766a2de4863a3e89316bc453144205aba4bdb690f1923b2ef"} Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.078424 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.080294 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" event={"ID":"0d112128-f54d-4045-88bc-3ddac499baab","Type":"ContainerStarted","Data":"3dfffe7b4b56bf22e7e6452fe16d940fce85e5072d1cef1e44c1506b2ff9eebf"} Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.080364 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" event={"ID":"0d112128-f54d-4045-88bc-3ddac499baab","Type":"ContainerStarted","Data":"307313599ba3c75a5f4ed215e0c5b9eeacb790594259db14e694480762abf713"} Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.080632 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.089667 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.098384 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" podStartSLOduration=3.098358273 podStartE2EDuration="3.098358273s" podCreationTimestamp="2026-02-20 00:12:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:12:33.094145968 +0000 UTC m=+237.288982027" watchObservedRunningTime="2026-02-20 00:12:33.098358273 +0000 UTC m=+237.293194322" Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.218390 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-86f5897b9d-75pfr" Feb 20 00:12:33 crc kubenswrapper[4750]: I0220 00:12:33.242322 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d5d6d49db-pzb99" podStartSLOduration=3.242298843 podStartE2EDuration="3.242298843s" podCreationTimestamp="2026-02-20 00:12:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:12:33.123929133 +0000 UTC m=+237.318765182" watchObservedRunningTime="2026-02-20 00:12:33.242298843 +0000 UTC m=+237.437134892" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.224380 4750 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.225602 4750 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.225885 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810" gracePeriod=15 Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.225943 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.225938 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa" gracePeriod=15 Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.225967 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe" gracePeriod=15 Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.226028 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539" gracePeriod=15 Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.225902 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd" gracePeriod=15 Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.228704 4750 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 00:12:45 crc kubenswrapper[4750]: E0220 00:12:45.229061 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229082 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 00:12:45 crc kubenswrapper[4750]: E0220 00:12:45.229103 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229141 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 00:12:45 crc kubenswrapper[4750]: E0220 00:12:45.229161 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229174 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 00:12:45 crc kubenswrapper[4750]: E0220 00:12:45.229199 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229211 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 00:12:45 crc kubenswrapper[4750]: E0220 00:12:45.229232 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229244 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 00:12:45 crc kubenswrapper[4750]: E0220 00:12:45.229260 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229272 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 00:12:45 crc kubenswrapper[4750]: E0220 00:12:45.229287 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229299 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229464 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229495 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229512 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229527 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229541 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.229557 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385548 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385603 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385626 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385646 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385738 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385854 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385899 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.385958 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.486919 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.486980 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487010 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487026 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487072 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487079 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487079 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487126 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487133 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487094 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487222 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487273 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487295 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487408 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487418 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:45 crc kubenswrapper[4750]: I0220 00:12:45.487468 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.154516 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.156475 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.157824 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd" exitCode=0 Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.157850 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539" exitCode=0 Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.157859 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa" exitCode=0 Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.157867 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe" exitCode=2 Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.157935 4750 scope.go:117] "RemoveContainer" containerID="5732116b2939bda09aa80322219ff2f41eda0e61854d07a5c1e90461c6d764c4" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.160663 4750 generic.go:334] "Generic (PLEG): container finished" podID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" containerID="6a3358f6e2340985b1d720644f8d3eabae89cca193a0f8786515bba5b41be3c9" exitCode=0 Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.160736 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bf2cf62d-cf5c-429a-b43c-228bae403a6e","Type":"ContainerDied","Data":"6a3358f6e2340985b1d720644f8d3eabae89cca193a0f8786515bba5b41be3c9"} Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.161924 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.162450 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.230545 4750 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.230607 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.562106 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:46 crc kubenswrapper[4750]: I0220 00:12:46.562824 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.167503 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.393517 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.394191 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.394623 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.395149 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.395580 4750 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.395621 4750 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.395888 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="200ms" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.584897 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.585637 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.586503 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.586944 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.597344 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="400ms" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.654544 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.655060 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.655310 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716227 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716304 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716328 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716374 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716424 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716499 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716685 4750 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716698 4750 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.716709 4750 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.817936 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-var-lock\") pod \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.818105 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kubelet-dir\") pod \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.818139 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-var-lock" (OuterVolumeSpecName: "var-lock") pod "bf2cf62d-cf5c-429a-b43c-228bae403a6e" (UID: "bf2cf62d-cf5c-429a-b43c-228bae403a6e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.818242 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kube-api-access\") pod \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\" (UID: \"bf2cf62d-cf5c-429a-b43c-228bae403a6e\") " Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.818210 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bf2cf62d-cf5c-429a-b43c-228bae403a6e" (UID: "bf2cf62d-cf5c-429a-b43c-228bae403a6e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.818617 4750 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.818650 4750 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.827355 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bf2cf62d-cf5c-429a-b43c-228bae403a6e" (UID: "bf2cf62d-cf5c-429a-b43c-228bae403a6e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:12:47 crc kubenswrapper[4750]: I0220 00:12:47.920222 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf2cf62d-cf5c-429a-b43c-228bae403a6e-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 00:12:47 crc kubenswrapper[4750]: E0220 00:12:47.998287 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="800ms" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.176569 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.177637 4750 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810" exitCode=0 Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.177785 4750 scope.go:117] "RemoveContainer" containerID="71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.177822 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.179229 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bf2cf62d-cf5c-429a-b43c-228bae403a6e","Type":"ContainerDied","Data":"6fa86484ac1fc81e50a71dced032b53e165c2ee36fa00e7716d80752e9520795"} Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.179267 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fa86484ac1fc81e50a71dced032b53e165c2ee36fa00e7716d80752e9520795" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.179320 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.192257 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.193085 4750 scope.go:117] "RemoveContainer" containerID="c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.193483 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.214225 4750 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.214333 4750 scope.go:117] "RemoveContainer" containerID="cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.215247 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.235001 4750 scope.go:117] "RemoveContainer" containerID="c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.248040 4750 scope.go:117] "RemoveContainer" containerID="345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.266475 4750 scope.go:117] "RemoveContainer" containerID="063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.281458 4750 scope.go:117] "RemoveContainer" containerID="71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.281763 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\": container with ID starting with 71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd not found: ID does not exist" containerID="71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.281789 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd"} err="failed to get container status \"71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\": rpc error: code = NotFound desc = could not find container \"71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd\": container with ID starting with 71a3e8ab0201a52d3bc83dcd7a79d9326ce700bfcd3d7e5881017f6aa9d12bcd not found: ID does not exist" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.281808 4750 scope.go:117] "RemoveContainer" containerID="c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.282101 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\": container with ID starting with c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539 not found: ID does not exist" containerID="c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.282180 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539"} err="failed to get container status \"c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\": rpc error: code = NotFound desc = could not find container \"c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539\": container with ID starting with c6a4d5cfead439725cce79e7c33661b17c5c8979914364bf6ba4ee6eea0ab539 not found: ID does not exist" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.282219 4750 scope.go:117] "RemoveContainer" containerID="cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.282594 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\": container with ID starting with cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa not found: ID does not exist" containerID="cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.282639 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa"} err="failed to get container status \"cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\": rpc error: code = NotFound desc = could not find container \"cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa\": container with ID starting with cfd58304efcfa3ea23b865b18cc5cc4db6f474d0e6a3ce54c8fa36289693e5fa not found: ID does not exist" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.282666 4750 scope.go:117] "RemoveContainer" containerID="c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.282943 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\": container with ID starting with c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe not found: ID does not exist" containerID="c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.282987 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe"} err="failed to get container status \"c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\": rpc error: code = NotFound desc = could not find container \"c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe\": container with ID starting with c5b1d4c4c23b9648451c434d7897d0dee397ee3a6638f5d1deda3cfd87ad6afe not found: ID does not exist" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.283016 4750 scope.go:117] "RemoveContainer" containerID="345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.284650 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\": container with ID starting with 345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810 not found: ID does not exist" containerID="345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.284692 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810"} err="failed to get container status \"345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\": rpc error: code = NotFound desc = could not find container \"345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810\": container with ID starting with 345dc2b03631c5341f56b69a14213975f90b86c523a253e1c1b3666826981810 not found: ID does not exist" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.284712 4750 scope.go:117] "RemoveContainer" containerID="063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.285095 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\": container with ID starting with 063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1 not found: ID does not exist" containerID="063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.285140 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1"} err="failed to get container status \"063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\": rpc error: code = NotFound desc = could not find container \"063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1\": container with ID starting with 063c06d6695ed7d9930511f13b6c3b5a32880d8afbd5d4efac5b6e0f2a8887f1 not found: ID does not exist" Feb 20 00:12:48 crc kubenswrapper[4750]: I0220 00:12:48.568543 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.717726 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:12:48Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:12:48Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:12:48Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T00:12:48Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.719253 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.719665 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.720301 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.720588 4750 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.720627 4750 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 00:12:48 crc kubenswrapper[4750]: E0220 00:12:48.798797 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="1.6s" Feb 20 00:12:50 crc kubenswrapper[4750]: E0220 00:12:50.267497 4750 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.194:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:50 crc kubenswrapper[4750]: I0220 00:12:50.268333 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:50 crc kubenswrapper[4750]: W0220 00:12:50.296524 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-08d10bcae24ef0c7fea8817a552bd3f9a2ed9f18faaff6e4086197caaa39557b WatchSource:0}: Error finding container 08d10bcae24ef0c7fea8817a552bd3f9a2ed9f18faaff6e4086197caaa39557b: Status 404 returned error can't find the container with id 08d10bcae24ef0c7fea8817a552bd3f9a2ed9f18faaff6e4086197caaa39557b Feb 20 00:12:50 crc kubenswrapper[4750]: E0220 00:12:50.300245 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895cc09a8d7853a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 00:12:50.29962681 +0000 UTC m=+254.494462859,LastTimestamp:2026-02-20 00:12:50.29962681 +0000 UTC m=+254.494462859,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 00:12:50 crc kubenswrapper[4750]: E0220 00:12:50.400426 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="3.2s" Feb 20 00:12:50 crc kubenswrapper[4750]: E0220 00:12:50.821931 4750 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.194:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.1895cc09a8d7853a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 00:12:50.29962681 +0000 UTC m=+254.494462859,LastTimestamp:2026-02-20 00:12:50.29962681 +0000 UTC m=+254.494462859,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 00:12:51 crc kubenswrapper[4750]: I0220 00:12:51.198277 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d396ab920c32b3bab86932e676fe6467681aac926f335b19225eca4ee27eb4b2"} Feb 20 00:12:51 crc kubenswrapper[4750]: I0220 00:12:51.198330 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"08d10bcae24ef0c7fea8817a552bd3f9a2ed9f18faaff6e4086197caaa39557b"} Feb 20 00:12:51 crc kubenswrapper[4750]: I0220 00:12:51.199100 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:51 crc kubenswrapper[4750]: E0220 00:12:51.199111 4750 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.194:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:12:53 crc kubenswrapper[4750]: E0220 00:12:53.602073 4750 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.194:6443: connect: connection refused" interval="6.4s" Feb 20 00:12:56 crc kubenswrapper[4750]: I0220 00:12:56.562901 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:57 crc kubenswrapper[4750]: I0220 00:12:57.559776 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:57 crc kubenswrapper[4750]: I0220 00:12:57.560915 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:57 crc kubenswrapper[4750]: I0220 00:12:57.585971 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:12:57 crc kubenswrapper[4750]: I0220 00:12:57.586016 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:12:57 crc kubenswrapper[4750]: E0220 00:12:57.586739 4750 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:57 crc kubenswrapper[4750]: I0220 00:12:57.587485 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:58 crc kubenswrapper[4750]: I0220 00:12:58.250157 4750 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="db8f546e361f940b48d9238736a4c6f38dcda76810aad863aea13812108a05e6" exitCode=0 Feb 20 00:12:58 crc kubenswrapper[4750]: I0220 00:12:58.250224 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"db8f546e361f940b48d9238736a4c6f38dcda76810aad863aea13812108a05e6"} Feb 20 00:12:58 crc kubenswrapper[4750]: I0220 00:12:58.250594 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6df4aa16cd94edb4fbd4818b46e2f4aecbfe42d290dcceb09e706dcb215f7830"} Feb 20 00:12:58 crc kubenswrapper[4750]: I0220 00:12:58.250841 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:12:58 crc kubenswrapper[4750]: I0220 00:12:58.250853 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:12:58 crc kubenswrapper[4750]: E0220 00:12:58.251391 4750 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:12:58 crc kubenswrapper[4750]: I0220 00:12:58.251496 4750 status_manager.go:851] "Failed to get status for pod" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.194:6443: connect: connection refused" Feb 20 00:12:59 crc kubenswrapper[4750]: I0220 00:12:59.271021 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7b26958931843dbd110fae7c2af2024f479c7f3fd1facd48034201ebd5960b38"} Feb 20 00:12:59 crc kubenswrapper[4750]: I0220 00:12:59.271070 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5ae93e13dfe619e28d9fb197013d50f81f65840aba5db3557c72c09acb0bbfcf"} Feb 20 00:12:59 crc kubenswrapper[4750]: I0220 00:12:59.271083 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a784f911de404f96a084614fd4cdd22b8c2227ee4fdef6e465b90c22dbafcc00"} Feb 20 00:12:59 crc kubenswrapper[4750]: I0220 00:12:59.276095 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 00:12:59 crc kubenswrapper[4750]: I0220 00:12:59.276159 4750 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082" exitCode=1 Feb 20 00:12:59 crc kubenswrapper[4750]: I0220 00:12:59.276185 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082"} Feb 20 00:12:59 crc kubenswrapper[4750]: I0220 00:12:59.276798 4750 scope.go:117] "RemoveContainer" containerID="08195b7fe772338ebabc2d0465e8e45069c1aa494ae49c195d7d8a4c175d5082" Feb 20 00:13:00 crc kubenswrapper[4750]: I0220 00:13:00.284965 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 00:13:00 crc kubenswrapper[4750]: I0220 00:13:00.285379 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"81c49dbea594c3e0a2892cb6cbde3164126818370876a2fc30dedd7db90dd5ab"} Feb 20 00:13:00 crc kubenswrapper[4750]: I0220 00:13:00.288571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fcd3ad3ada84017c5f26f68be27f12a910ad60ddb7c524564d745f43f7825296"} Feb 20 00:13:00 crc kubenswrapper[4750]: I0220 00:13:00.288592 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4b16c95f676756cb8d9ceafa3117dc95961d88abff9f748ee8f2bd6aa2c7a226"} Feb 20 00:13:00 crc kubenswrapper[4750]: I0220 00:13:00.288760 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:13:00 crc kubenswrapper[4750]: I0220 00:13:00.288778 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:13:00 crc kubenswrapper[4750]: I0220 00:13:00.288992 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:13:02 crc kubenswrapper[4750]: I0220 00:13:02.587782 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:13:02 crc kubenswrapper[4750]: I0220 00:13:02.587827 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:13:02 crc kubenswrapper[4750]: I0220 00:13:02.594153 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:13:03 crc kubenswrapper[4750]: I0220 00:13:03.879866 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:13:03 crc kubenswrapper[4750]: I0220 00:13:03.886288 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:13:04 crc kubenswrapper[4750]: I0220 00:13:04.320264 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:13:05 crc kubenswrapper[4750]: I0220 00:13:05.318314 4750 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:13:06 crc kubenswrapper[4750]: I0220 00:13:06.333461 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:13:06 crc kubenswrapper[4750]: I0220 00:13:06.335014 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:13:06 crc kubenswrapper[4750]: I0220 00:13:06.342345 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:13:06 crc kubenswrapper[4750]: I0220 00:13:06.590303 4750 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d91655fd-50ff-44a9-bdf4-3f65204e5fa9" Feb 20 00:13:07 crc kubenswrapper[4750]: I0220 00:13:07.340560 4750 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:13:07 crc kubenswrapper[4750]: I0220 00:13:07.340603 4750 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3b6b1e32-a6a0-4ed3-b542-9d898afde9dc" Feb 20 00:13:07 crc kubenswrapper[4750]: I0220 00:13:07.346476 4750 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d91655fd-50ff-44a9-bdf4-3f65204e5fa9" Feb 20 00:13:15 crc kubenswrapper[4750]: I0220 00:13:15.647406 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 00:13:15 crc kubenswrapper[4750]: I0220 00:13:15.658164 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 00:13:15 crc kubenswrapper[4750]: I0220 00:13:15.928591 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 00:13:15 crc kubenswrapper[4750]: I0220 00:13:15.979245 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 00:13:16 crc kubenswrapper[4750]: I0220 00:13:16.207044 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 00:13:16 crc kubenswrapper[4750]: I0220 00:13:16.657506 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 00:13:16 crc kubenswrapper[4750]: I0220 00:13:16.683800 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 00:13:16 crc kubenswrapper[4750]: I0220 00:13:16.824921 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.181495 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.368758 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.448377 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.506947 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.631313 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.737633 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.738403 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.867960 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 00:13:17 crc kubenswrapper[4750]: I0220 00:13:17.976263 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.031702 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.084440 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.105513 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.176598 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.646953 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.674248 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.831839 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 00:13:18 crc kubenswrapper[4750]: I0220 00:13:18.854081 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.094558 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.122591 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.210431 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.243265 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.338964 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.434218 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.480098 4750 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.494551 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.625886 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.708012 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.765688 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.855215 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.904556 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 00:13:19 crc kubenswrapper[4750]: I0220 00:13:19.918696 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.007754 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.031600 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.167825 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.174080 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.242650 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.280022 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.312152 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.329599 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.369767 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.385997 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.411256 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.437143 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.503911 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.508692 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.540426 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.688794 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.720054 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.778717 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.811288 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.850573 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.969899 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 00:13:20 crc kubenswrapper[4750]: I0220 00:13:20.978110 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.116888 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.201256 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.208326 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.208480 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.268379 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.284744 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.322442 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.390846 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.429648 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.513624 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.619560 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.691699 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.814976 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.828647 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.865944 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 00:13:21 crc kubenswrapper[4750]: I0220 00:13:21.914502 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.019966 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.054139 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.134146 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.249499 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.267992 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.284830 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.307758 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.426987 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.433306 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.495645 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.569847 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.671907 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.680239 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.780551 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.782460 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.838016 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.853693 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.898686 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.960761 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 00:13:22 crc kubenswrapper[4750]: I0220 00:13:22.969564 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.076053 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.252369 4750 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.256908 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.359468 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.410920 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.578078 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.605538 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.668791 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.671919 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.686904 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.710742 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.915675 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 00:13:23 crc kubenswrapper[4750]: I0220 00:13:23.957298 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.054710 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.112595 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.160685 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.176771 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.191673 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.196185 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.255407 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.272090 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.343722 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.531998 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.556778 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.557568 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.583707 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.589556 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.642244 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.740789 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.741072 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.765353 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.902759 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 00:13:24 crc kubenswrapper[4750]: I0220 00:13:24.938403 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.008020 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.111998 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.178341 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.196019 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.247501 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.314499 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.353904 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.394482 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.470586 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.484414 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.653850 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.770026 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.771496 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.780936 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.782351 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.815940 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.822775 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 00:13:25 crc kubenswrapper[4750]: I0220 00:13:25.937386 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.029101 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.051867 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.082567 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.142755 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.143591 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.232569 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.292142 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.541638 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.565018 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.601049 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.634776 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.634952 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.708660 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.903047 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.908877 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.914959 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.947183 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.961997 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 00:13:26 crc kubenswrapper[4750]: I0220 00:13:26.985819 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.131326 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.149032 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.252515 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.265634 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.298372 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.306682 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.325018 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.329882 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.423371 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.467582 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.478684 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.486339 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.511606 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.584477 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.601252 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.613720 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.669173 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.861379 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.953536 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 00:13:27 crc kubenswrapper[4750]: I0220 00:13:27.962761 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.113190 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.151983 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.270570 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.368922 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.456469 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.457913 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.576576 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.588528 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.655187 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 00:13:28 crc kubenswrapper[4750]: I0220 00:13:28.882998 4750 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.054463 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.127250 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.147428 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.190460 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.224741 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.233708 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.260399 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.288961 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.290648 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.298519 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.313302 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.375161 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.397209 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.432695 4750 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.437329 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.437388 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.443574 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.447949 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.456702 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.456678125 podStartE2EDuration="24.456678125s" podCreationTimestamp="2026-02-20 00:13:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:13:29.450901416 +0000 UTC m=+293.645737465" watchObservedRunningTime="2026-02-20 00:13:29.456678125 +0000 UTC m=+293.651514174" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.481168 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.503625 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.564533 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.674578 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.689623 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.698516 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.738054 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.774465 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.807004 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.887326 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 00:13:29 crc kubenswrapper[4750]: I0220 00:13:29.939968 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.013519 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.076677 4750 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.100494 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.123923 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.171275 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.336483 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.484337 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.551081 4750 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.598957 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.764828 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 00:13:30 crc kubenswrapper[4750]: I0220 00:13:30.937992 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.151387 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.170855 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.171947 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.311004 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.356028 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.612307 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.620153 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.732658 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.786351 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.822332 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.851502 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 00:13:31 crc kubenswrapper[4750]: I0220 00:13:31.923561 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.126213 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.552711 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.822956 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62qrs"] Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.834673 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rgp2g"] Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.835047 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rgp2g" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="registry-server" containerID="cri-o://c2649448706f267d1e7a8cc5fef1d582664d265352c8ffc1ad9ec8cd50e43b3e" gracePeriod=30 Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.855461 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gksdw"] Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.855828 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" podUID="55531783-59c5-4d74-b4af-d9e89c0a8421" containerName="marketplace-operator" containerID="cri-o://6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a" gracePeriod=30 Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.870511 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2vd9"] Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.870911 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z2vd9" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="registry-server" containerID="cri-o://67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b" gracePeriod=30 Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.878320 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sll9r"] Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.878889 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sll9r" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="registry-server" containerID="cri-o://d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3" gracePeriod=30 Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.890723 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.899354 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dvbwr"] Feb 20 00:13:32 crc kubenswrapper[4750]: E0220 00:13:32.899627 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" containerName="installer" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.899647 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" containerName="installer" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.899780 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2cf62d-cf5c-429a-b43c-228bae403a6e" containerName="installer" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.900230 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.908110 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dvbwr"] Feb 20 00:13:32 crc kubenswrapper[4750]: I0220 00:13:32.954950 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.054175 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnhhn\" (UniqueName: \"kubernetes.io/projected/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-kube-api-access-tnhhn\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.054249 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.054312 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.156301 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnhhn\" (UniqueName: \"kubernetes.io/projected/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-kube-api-access-tnhhn\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.156385 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.156442 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.158250 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.164461 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.176579 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnhhn\" (UniqueName: \"kubernetes.io/projected/f26d2d81-da99-4a29-9f9f-f975eb3a75f3-kube-api-access-tnhhn\") pod \"marketplace-operator-79b997595-dvbwr\" (UID: \"f26d2d81-da99-4a29-9f9f-f975eb3a75f3\") " pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.198822 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.241804 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.243645 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.299513 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.326823 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.347804 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.352462 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459483 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-trusted-ca\") pod \"55531783-59c5-4d74-b4af-d9e89c0a8421\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459547 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-utilities\") pod \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459635 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlgqf\" (UniqueName: \"kubernetes.io/projected/295b160a-bc39-40be-822d-fc4438895480-kube-api-access-wlgqf\") pod \"295b160a-bc39-40be-822d-fc4438895480\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459700 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zck6g\" (UniqueName: \"kubernetes.io/projected/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-kube-api-access-zck6g\") pod \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459737 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-catalog-content\") pod \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\" (UID: \"0e3cf328-65f5-417b-8bb4-35ff194e8cc4\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459772 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgjnd\" (UniqueName: \"kubernetes.io/projected/55531783-59c5-4d74-b4af-d9e89c0a8421-kube-api-access-tgjnd\") pod \"55531783-59c5-4d74-b4af-d9e89c0a8421\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459810 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-utilities\") pod \"295b160a-bc39-40be-822d-fc4438895480\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459852 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-operator-metrics\") pod \"55531783-59c5-4d74-b4af-d9e89c0a8421\" (UID: \"55531783-59c5-4d74-b4af-d9e89c0a8421\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.459989 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-catalog-content\") pod \"295b160a-bc39-40be-822d-fc4438895480\" (UID: \"295b160a-bc39-40be-822d-fc4438895480\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.460035 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "55531783-59c5-4d74-b4af-d9e89c0a8421" (UID: "55531783-59c5-4d74-b4af-d9e89c0a8421"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.460245 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.461148 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-utilities" (OuterVolumeSpecName: "utilities") pod "0e3cf328-65f5-417b-8bb4-35ff194e8cc4" (UID: "0e3cf328-65f5-417b-8bb4-35ff194e8cc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.462207 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-utilities" (OuterVolumeSpecName: "utilities") pod "295b160a-bc39-40be-822d-fc4438895480" (UID: "295b160a-bc39-40be-822d-fc4438895480"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.473407 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-kube-api-access-zck6g" (OuterVolumeSpecName: "kube-api-access-zck6g") pod "0e3cf328-65f5-417b-8bb4-35ff194e8cc4" (UID: "0e3cf328-65f5-417b-8bb4-35ff194e8cc4"). InnerVolumeSpecName "kube-api-access-zck6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.473468 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55531783-59c5-4d74-b4af-d9e89c0a8421-kube-api-access-tgjnd" (OuterVolumeSpecName: "kube-api-access-tgjnd") pod "55531783-59c5-4d74-b4af-d9e89c0a8421" (UID: "55531783-59c5-4d74-b4af-d9e89c0a8421"). InnerVolumeSpecName "kube-api-access-tgjnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.473465 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295b160a-bc39-40be-822d-fc4438895480-kube-api-access-wlgqf" (OuterVolumeSpecName: "kube-api-access-wlgqf") pod "295b160a-bc39-40be-822d-fc4438895480" (UID: "295b160a-bc39-40be-822d-fc4438895480"). InnerVolumeSpecName "kube-api-access-wlgqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.473687 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "55531783-59c5-4d74-b4af-d9e89c0a8421" (UID: "55531783-59c5-4d74-b4af-d9e89c0a8421"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.486884 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e3cf328-65f5-417b-8bb4-35ff194e8cc4" (UID: "0e3cf328-65f5-417b-8bb4-35ff194e8cc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.512385 4750 generic.go:334] "Generic (PLEG): container finished" podID="55531783-59c5-4d74-b4af-d9e89c0a8421" containerID="6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a" exitCode=0 Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.512460 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" event={"ID":"55531783-59c5-4d74-b4af-d9e89c0a8421","Type":"ContainerDied","Data":"6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a"} Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.512492 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" event={"ID":"55531783-59c5-4d74-b4af-d9e89c0a8421","Type":"ContainerDied","Data":"644278be123dee14b92de7c080ace0415aa47238664a2e3dfe770ea435462a61"} Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.512510 4750 scope.go:117] "RemoveContainer" containerID="6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.512649 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gksdw" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.515421 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgp2g" event={"ID":"9ce86efc-8081-4418-ac52-06ce3348bc69","Type":"ContainerDied","Data":"c2649448706f267d1e7a8cc5fef1d582664d265352c8ffc1ad9ec8cd50e43b3e"} Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.515514 4750 generic.go:334] "Generic (PLEG): container finished" podID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerID="c2649448706f267d1e7a8cc5fef1d582664d265352c8ffc1ad9ec8cd50e43b3e" exitCode=0 Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.517686 4750 generic.go:334] "Generic (PLEG): container finished" podID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerID="67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b" exitCode=0 Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.517727 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2vd9" event={"ID":"0e3cf328-65f5-417b-8bb4-35ff194e8cc4","Type":"ContainerDied","Data":"67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b"} Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.517745 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2vd9" event={"ID":"0e3cf328-65f5-417b-8bb4-35ff194e8cc4","Type":"ContainerDied","Data":"a29a1c79128721128c764e69281513c8874538d70f672b3cc47820150bfb3f07"} Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.517824 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2vd9" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.527685 4750 generic.go:334] "Generic (PLEG): container finished" podID="295b160a-bc39-40be-822d-fc4438895480" containerID="d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3" exitCode=0 Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.527895 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-62qrs" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="registry-server" containerID="cri-o://03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a" gracePeriod=30 Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.528203 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sll9r" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.528419 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sll9r" event={"ID":"295b160a-bc39-40be-822d-fc4438895480","Type":"ContainerDied","Data":"d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3"} Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.528452 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sll9r" event={"ID":"295b160a-bc39-40be-822d-fc4438895480","Type":"ContainerDied","Data":"e2d36dad879b2e53b716779d32d770d3d42cc7123a33f12877e1850bc7f131cf"} Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.550819 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gksdw"] Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.555724 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gksdw"] Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561033 4750 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/55531783-59c5-4d74-b4af-d9e89c0a8421-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561062 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561074 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlgqf\" (UniqueName: \"kubernetes.io/projected/295b160a-bc39-40be-822d-fc4438895480-kube-api-access-wlgqf\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561083 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zck6g\" (UniqueName: \"kubernetes.io/projected/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-kube-api-access-zck6g\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561091 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3cf328-65f5-417b-8bb4-35ff194e8cc4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561099 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgjnd\" (UniqueName: \"kubernetes.io/projected/55531783-59c5-4d74-b4af-d9e89c0a8421-kube-api-access-tgjnd\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561107 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.561967 4750 scope.go:117] "RemoveContainer" containerID="6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a" Feb 20 00:13:33 crc kubenswrapper[4750]: E0220 00:13:33.563161 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a\": container with ID starting with 6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a not found: ID does not exist" containerID="6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.563235 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a"} err="failed to get container status \"6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a\": rpc error: code = NotFound desc = could not find container \"6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a\": container with ID starting with 6d725a78b54afef3e1e91d5fc04b58c6328add2c8d170e1f7c86491fece5f36a not found: ID does not exist" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.563269 4750 scope.go:117] "RemoveContainer" containerID="67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.567005 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2vd9"] Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.572500 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2vd9"] Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.576526 4750 scope.go:117] "RemoveContainer" containerID="de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.592325 4750 scope.go:117] "RemoveContainer" containerID="9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.613877 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "295b160a-bc39-40be-822d-fc4438895480" (UID: "295b160a-bc39-40be-822d-fc4438895480"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.656465 4750 scope.go:117] "RemoveContainer" containerID="67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b" Feb 20 00:13:33 crc kubenswrapper[4750]: E0220 00:13:33.657257 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b\": container with ID starting with 67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b not found: ID does not exist" containerID="67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.657304 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b"} err="failed to get container status \"67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b\": rpc error: code = NotFound desc = could not find container \"67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b\": container with ID starting with 67ca312eae972966cbf9302cb236b4a979e80c79a8101c81b2c6ef24774e479b not found: ID does not exist" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.657332 4750 scope.go:117] "RemoveContainer" containerID="de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698" Feb 20 00:13:33 crc kubenswrapper[4750]: E0220 00:13:33.657745 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698\": container with ID starting with de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698 not found: ID does not exist" containerID="de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.657766 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698"} err="failed to get container status \"de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698\": rpc error: code = NotFound desc = could not find container \"de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698\": container with ID starting with de9f4aa24c4ced907a03fec4112c32c964dee6375622199697e0344e2df26698 not found: ID does not exist" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.657779 4750 scope.go:117] "RemoveContainer" containerID="9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271" Feb 20 00:13:33 crc kubenswrapper[4750]: E0220 00:13:33.657959 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271\": container with ID starting with 9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271 not found: ID does not exist" containerID="9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.657978 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271"} err="failed to get container status \"9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271\": rpc error: code = NotFound desc = could not find container \"9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271\": container with ID starting with 9862720debbfbad9b3e625577e4301e1d74c0366841e46433e030a383ade1271 not found: ID does not exist" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.657991 4750 scope.go:117] "RemoveContainer" containerID="d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.661718 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/295b160a-bc39-40be-822d-fc4438895480-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.675920 4750 scope.go:117] "RemoveContainer" containerID="b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.686899 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dvbwr"] Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.720818 4750 scope.go:117] "RemoveContainer" containerID="f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.732516 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.743639 4750 scope.go:117] "RemoveContainer" containerID="d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3" Feb 20 00:13:33 crc kubenswrapper[4750]: E0220 00:13:33.744085 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3\": container with ID starting with d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3 not found: ID does not exist" containerID="d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.744145 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3"} err="failed to get container status \"d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3\": rpc error: code = NotFound desc = could not find container \"d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3\": container with ID starting with d8252574d01308b341f5df4accacc6bd8338df82eb336ecc9d3ce58a1ad9e7f3 not found: ID does not exist" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.744174 4750 scope.go:117] "RemoveContainer" containerID="b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93" Feb 20 00:13:33 crc kubenswrapper[4750]: E0220 00:13:33.745043 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93\": container with ID starting with b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93 not found: ID does not exist" containerID="b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.745085 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93"} err="failed to get container status \"b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93\": rpc error: code = NotFound desc = could not find container \"b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93\": container with ID starting with b790ba25c1ab3e6c63a7854aef64698c1722c6cd644e80f7a63bd07ddb856b93 not found: ID does not exist" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.745120 4750 scope.go:117] "RemoveContainer" containerID="f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb" Feb 20 00:13:33 crc kubenswrapper[4750]: E0220 00:13:33.747063 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb\": container with ID starting with f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb not found: ID does not exist" containerID="f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.747109 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb"} err="failed to get container status \"f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb\": rpc error: code = NotFound desc = could not find container \"f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb\": container with ID starting with f1eebd730248bff96a8ef4820e79c331d05ca74e6dddf9e9eea21ac30bc9fbfb not found: ID does not exist" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.861607 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sll9r"] Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.864314 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-utilities\") pod \"9ce86efc-8081-4418-ac52-06ce3348bc69\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.864366 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-catalog-content\") pod \"9ce86efc-8081-4418-ac52-06ce3348bc69\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.864411 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qqzn\" (UniqueName: \"kubernetes.io/projected/9ce86efc-8081-4418-ac52-06ce3348bc69-kube-api-access-5qqzn\") pod \"9ce86efc-8081-4418-ac52-06ce3348bc69\" (UID: \"9ce86efc-8081-4418-ac52-06ce3348bc69\") " Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.865169 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-utilities" (OuterVolumeSpecName: "utilities") pod "9ce86efc-8081-4418-ac52-06ce3348bc69" (UID: "9ce86efc-8081-4418-ac52-06ce3348bc69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.865567 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.866700 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sll9r"] Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.868943 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce86efc-8081-4418-ac52-06ce3348bc69-kube-api-access-5qqzn" (OuterVolumeSpecName: "kube-api-access-5qqzn") pod "9ce86efc-8081-4418-ac52-06ce3348bc69" (UID: "9ce86efc-8081-4418-ac52-06ce3348bc69"). InnerVolumeSpecName "kube-api-access-5qqzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.912510 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ce86efc-8081-4418-ac52-06ce3348bc69" (UID: "9ce86efc-8081-4418-ac52-06ce3348bc69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.925121 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.967095 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qqzn\" (UniqueName: \"kubernetes.io/projected/9ce86efc-8081-4418-ac52-06ce3348bc69-kube-api-access-5qqzn\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:33 crc kubenswrapper[4750]: I0220 00:13:33.967145 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ce86efc-8081-4418-ac52-06ce3348bc69-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.068751 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-catalog-content\") pod \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.068915 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vklkw\" (UniqueName: \"kubernetes.io/projected/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-kube-api-access-vklkw\") pod \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.069025 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-utilities\") pod \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\" (UID: \"3a63cf32-f756-45bb-ab8d-1e3aff3d2165\") " Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.070547 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-utilities" (OuterVolumeSpecName: "utilities") pod "3a63cf32-f756-45bb-ab8d-1e3aff3d2165" (UID: "3a63cf32-f756-45bb-ab8d-1e3aff3d2165"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.073333 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-kube-api-access-vklkw" (OuterVolumeSpecName: "kube-api-access-vklkw") pod "3a63cf32-f756-45bb-ab8d-1e3aff3d2165" (UID: "3a63cf32-f756-45bb-ab8d-1e3aff3d2165"). InnerVolumeSpecName "kube-api-access-vklkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.114763 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a63cf32-f756-45bb-ab8d-1e3aff3d2165" (UID: "3a63cf32-f756-45bb-ab8d-1e3aff3d2165"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.171084 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vklkw\" (UniqueName: \"kubernetes.io/projected/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-kube-api-access-vklkw\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.171151 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.171164 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a63cf32-f756-45bb-ab8d-1e3aff3d2165-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.538292 4750 generic.go:334] "Generic (PLEG): container finished" podID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerID="03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a" exitCode=0 Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.538392 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62qrs" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.538401 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62qrs" event={"ID":"3a63cf32-f756-45bb-ab8d-1e3aff3d2165","Type":"ContainerDied","Data":"03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a"} Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.538927 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62qrs" event={"ID":"3a63cf32-f756-45bb-ab8d-1e3aff3d2165","Type":"ContainerDied","Data":"7579ca9463da1cd7d5bed2a0949beefc7ac468dadc690f774c9fcc7728e503c0"} Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.538961 4750 scope.go:117] "RemoveContainer" containerID="03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.540643 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" event={"ID":"f26d2d81-da99-4a29-9f9f-f975eb3a75f3","Type":"ContainerStarted","Data":"184292193adb9a1eb24b8aee8bfa7530fde8c16696c4e7362acbdc0c35ad87ed"} Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.540928 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.541737 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" event={"ID":"f26d2d81-da99-4a29-9f9f-f975eb3a75f3","Type":"ContainerStarted","Data":"1af439cfc6bcd85eb3c9a9e9058e1ef16e22507d4a95da9998a4d9246d8d91c2"} Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.547406 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rgp2g" event={"ID":"9ce86efc-8081-4418-ac52-06ce3348bc69","Type":"ContainerDied","Data":"30cda0254c35d0de23abc101c09b14a4427574f52e0ba4ff3fd6bec3575eb845"} Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.547499 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rgp2g" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.550287 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.567229 4750 scope.go:117] "RemoveContainer" containerID="091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.573480 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" path="/var/lib/kubelet/pods/0e3cf328-65f5-417b-8bb4-35ff194e8cc4/volumes" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.575254 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295b160a-bc39-40be-822d-fc4438895480" path="/var/lib/kubelet/pods/295b160a-bc39-40be-822d-fc4438895480/volumes" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.577312 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55531783-59c5-4d74-b4af-d9e89c0a8421" path="/var/lib/kubelet/pods/55531783-59c5-4d74-b4af-d9e89c0a8421/volumes" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.586079 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dvbwr" podStartSLOduration=2.586055853 podStartE2EDuration="2.586055853s" podCreationTimestamp="2026-02-20 00:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:13:34.583263646 +0000 UTC m=+298.778099735" watchObservedRunningTime="2026-02-20 00:13:34.586055853 +0000 UTC m=+298.780891912" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.613872 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rgp2g"] Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.616209 4750 scope.go:117] "RemoveContainer" containerID="2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.621631 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rgp2g"] Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.635371 4750 scope.go:117] "RemoveContainer" containerID="03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a" Feb 20 00:13:34 crc kubenswrapper[4750]: E0220 00:13:34.635987 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a\": container with ID starting with 03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a not found: ID does not exist" containerID="03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.636050 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a"} err="failed to get container status \"03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a\": rpc error: code = NotFound desc = could not find container \"03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a\": container with ID starting with 03a759ac81c99e4df3e6ce6c11876aa995aed64cc93c22df00c7ee33c4daaa7a not found: ID does not exist" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.636075 4750 scope.go:117] "RemoveContainer" containerID="091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733" Feb 20 00:13:34 crc kubenswrapper[4750]: E0220 00:13:34.638509 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733\": container with ID starting with 091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733 not found: ID does not exist" containerID="091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.638559 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733"} err="failed to get container status \"091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733\": rpc error: code = NotFound desc = could not find container \"091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733\": container with ID starting with 091a48f122f0cadd5d906438d2282aeb673494b82a9f3ea3609498589b6bb733 not found: ID does not exist" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.638590 4750 scope.go:117] "RemoveContainer" containerID="2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.638934 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62qrs"] Feb 20 00:13:34 crc kubenswrapper[4750]: E0220 00:13:34.639788 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317\": container with ID starting with 2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317 not found: ID does not exist" containerID="2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.639822 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317"} err="failed to get container status \"2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317\": rpc error: code = NotFound desc = could not find container \"2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317\": container with ID starting with 2b3457e765a6b0b084cadf97d3e61ece8c41e357cef2858f76b0294083ef8317 not found: ID does not exist" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.639846 4750 scope.go:117] "RemoveContainer" containerID="c2649448706f267d1e7a8cc5fef1d582664d265352c8ffc1ad9ec8cd50e43b3e" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.647182 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-62qrs"] Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.667878 4750 scope.go:117] "RemoveContainer" containerID="55250ea9abcac1cf7012db718d4261da7c430216e4687651e136ca16dea6bb7a" Feb 20 00:13:34 crc kubenswrapper[4750]: I0220 00:13:34.697312 4750 scope.go:117] "RemoveContainer" containerID="b43e1a9380ab0c28a54b7d6de786e862061af00435e5a5bec32c2d73d7efac4c" Feb 20 00:13:36 crc kubenswrapper[4750]: I0220 00:13:36.320063 4750 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 20 00:13:36 crc kubenswrapper[4750]: I0220 00:13:36.572690 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" path="/var/lib/kubelet/pods/3a63cf32-f756-45bb-ab8d-1e3aff3d2165/volumes" Feb 20 00:13:36 crc kubenswrapper[4750]: I0220 00:13:36.573321 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" path="/var/lib/kubelet/pods/9ce86efc-8081-4418-ac52-06ce3348bc69/volumes" Feb 20 00:13:39 crc kubenswrapper[4750]: I0220 00:13:39.301676 4750 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 00:13:39 crc kubenswrapper[4750]: I0220 00:13:39.303330 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d396ab920c32b3bab86932e676fe6467681aac926f335b19225eca4ee27eb4b2" gracePeriod=5 Feb 20 00:13:44 crc kubenswrapper[4750]: I0220 00:13:44.606065 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 00:13:44 crc kubenswrapper[4750]: I0220 00:13:44.606518 4750 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d396ab920c32b3bab86932e676fe6467681aac926f335b19225eca4ee27eb4b2" exitCode=137 Feb 20 00:13:44 crc kubenswrapper[4750]: I0220 00:13:44.898828 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 00:13:44 crc kubenswrapper[4750]: I0220 00:13:44.898911 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007290 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007409 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007455 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007499 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007641 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007860 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007931 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.007976 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.008103 4750 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.008300 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.015337 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.110215 4750 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.110636 4750 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.110654 4750 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.110672 4750 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.613027 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.614888 4750 scope.go:117] "RemoveContainer" containerID="d396ab920c32b3bab86932e676fe6467681aac926f335b19225eca4ee27eb4b2" Feb 20 00:13:45 crc kubenswrapper[4750]: I0220 00:13:45.614965 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 00:13:46 crc kubenswrapper[4750]: I0220 00:13:46.565158 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.808496 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m6jfb"] Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809286 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809301 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809316 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809324 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809337 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809346 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809358 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809367 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809382 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55531783-59c5-4d74-b4af-d9e89c0a8421" containerName="marketplace-operator" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809392 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="55531783-59c5-4d74-b4af-d9e89c0a8421" containerName="marketplace-operator" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809406 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809414 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809424 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809432 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809445 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809453 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809462 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809471 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809485 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809493 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809507 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809516 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809526 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809534 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="extract-utilities" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809545 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809553 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: E0220 00:13:55.809563 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809572 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="extract-content" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809677 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="55531783-59c5-4d74-b4af-d9e89c0a8421" containerName="marketplace-operator" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809691 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a63cf32-f756-45bb-ab8d-1e3aff3d2165" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809706 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ce86efc-8081-4418-ac52-06ce3348bc69" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809716 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e3cf328-65f5-417b-8bb4-35ff194e8cc4" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809730 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.809741 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="295b160a-bc39-40be-822d-fc4438895480" containerName="registry-server" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.810604 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.816226 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.821324 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m6jfb"] Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.953532 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-catalog-content\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.953656 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h68td\" (UniqueName: \"kubernetes.io/projected/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-kube-api-access-h68td\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:55 crc kubenswrapper[4750]: I0220 00:13:55.953701 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-utilities\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.002810 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pf4pd"] Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.003697 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.005399 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.011717 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pf4pd"] Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.055225 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-utilities\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.055292 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-catalog-content\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.055359 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h68td\" (UniqueName: \"kubernetes.io/projected/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-kube-api-access-h68td\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.056139 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-utilities\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.056380 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-catalog-content\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.075643 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h68td\" (UniqueName: \"kubernetes.io/projected/abac413a-7c83-4f73-8c5b-b1c0d50d5ebd-kube-api-access-h68td\") pod \"redhat-operators-m6jfb\" (UID: \"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd\") " pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.135595 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.156077 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-catalog-content\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.156149 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-utilities\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.156187 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8fl5\" (UniqueName: \"kubernetes.io/projected/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-kube-api-access-g8fl5\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.256929 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-catalog-content\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.257330 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-utilities\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.257358 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8fl5\" (UniqueName: \"kubernetes.io/projected/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-kube-api-access-g8fl5\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.257617 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-catalog-content\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.258058 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-utilities\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.277637 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8fl5\" (UniqueName: \"kubernetes.io/projected/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-kube-api-access-g8fl5\") pod \"redhat-marketplace-pf4pd\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.323674 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.591565 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m6jfb"] Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.678903 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6jfb" event={"ID":"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd","Type":"ContainerStarted","Data":"3eb2abab8dedafa735e5d028b2a28a9f41c3c276b7294e065b8ee1bbb4d263b2"} Feb 20 00:13:56 crc kubenswrapper[4750]: I0220 00:13:56.730788 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pf4pd"] Feb 20 00:13:56 crc kubenswrapper[4750]: W0220 00:13:56.818722 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8a980d8_c735_4920_a9e3_ee6eb98b00b1.slice/crio-390e82ed07e0638b7e54cd0b81d144096e50037cb2f9db3dc0999d51ad99c5f7 WatchSource:0}: Error finding container 390e82ed07e0638b7e54cd0b81d144096e50037cb2f9db3dc0999d51ad99c5f7: Status 404 returned error can't find the container with id 390e82ed07e0638b7e54cd0b81d144096e50037cb2f9db3dc0999d51ad99c5f7 Feb 20 00:13:57 crc kubenswrapper[4750]: I0220 00:13:57.686904 4750 generic.go:334] "Generic (PLEG): container finished" podID="abac413a-7c83-4f73-8c5b-b1c0d50d5ebd" containerID="f3ac829dae1c066590abf8540bf4b05e4224245946386e020867bd04cec6e1e7" exitCode=0 Feb 20 00:13:57 crc kubenswrapper[4750]: I0220 00:13:57.687014 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6jfb" event={"ID":"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd","Type":"ContainerDied","Data":"f3ac829dae1c066590abf8540bf4b05e4224245946386e020867bd04cec6e1e7"} Feb 20 00:13:57 crc kubenswrapper[4750]: I0220 00:13:57.689515 4750 generic.go:334] "Generic (PLEG): container finished" podID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerID="a2146a4334faa3fb46dc02bb85b8b98702ef1196aa4d3ab4db4a283a987f8ef2" exitCode=0 Feb 20 00:13:57 crc kubenswrapper[4750]: I0220 00:13:57.689576 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pf4pd" event={"ID":"a8a980d8-c735-4920-a9e3-ee6eb98b00b1","Type":"ContainerDied","Data":"a2146a4334faa3fb46dc02bb85b8b98702ef1196aa4d3ab4db4a283a987f8ef2"} Feb 20 00:13:57 crc kubenswrapper[4750]: I0220 00:13:57.689610 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pf4pd" event={"ID":"a8a980d8-c735-4920-a9e3-ee6eb98b00b1","Type":"ContainerStarted","Data":"390e82ed07e0638b7e54cd0b81d144096e50037cb2f9db3dc0999d51ad99c5f7"} Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.206270 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5px4g"] Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.208313 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.218975 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5px4g"] Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.219144 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.282136 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fc4f50-400a-417e-9538-ef3dd5a66038-catalog-content\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.282183 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fc4f50-400a-417e-9538-ef3dd5a66038-utilities\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.282213 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwqrc\" (UniqueName: \"kubernetes.io/projected/44fc4f50-400a-417e-9538-ef3dd5a66038-kube-api-access-pwqrc\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.384596 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fc4f50-400a-417e-9538-ef3dd5a66038-catalog-content\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.384901 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fc4f50-400a-417e-9538-ef3dd5a66038-utilities\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.385188 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwqrc\" (UniqueName: \"kubernetes.io/projected/44fc4f50-400a-417e-9538-ef3dd5a66038-kube-api-access-pwqrc\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.385645 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44fc4f50-400a-417e-9538-ef3dd5a66038-utilities\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.386046 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44fc4f50-400a-417e-9538-ef3dd5a66038-catalog-content\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.416082 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwqrc\" (UniqueName: \"kubernetes.io/projected/44fc4f50-400a-417e-9538-ef3dd5a66038-kube-api-access-pwqrc\") pod \"certified-operators-5px4g\" (UID: \"44fc4f50-400a-417e-9538-ef3dd5a66038\") " pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.423360 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mfjrj"] Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.428750 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mfjrj"] Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.428874 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.431059 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.486607 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfe5c0b-cde1-415e-846f-73947bfa62b4-utilities\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.486696 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss9dl\" (UniqueName: \"kubernetes.io/projected/2bfe5c0b-cde1-415e-846f-73947bfa62b4-kube-api-access-ss9dl\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.486732 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfe5c0b-cde1-415e-846f-73947bfa62b4-catalog-content\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.545069 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.588292 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss9dl\" (UniqueName: \"kubernetes.io/projected/2bfe5c0b-cde1-415e-846f-73947bfa62b4-kube-api-access-ss9dl\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.588550 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfe5c0b-cde1-415e-846f-73947bfa62b4-catalog-content\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.588592 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfe5c0b-cde1-415e-846f-73947bfa62b4-utilities\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.589031 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfe5c0b-cde1-415e-846f-73947bfa62b4-utilities\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.589378 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfe5c0b-cde1-415e-846f-73947bfa62b4-catalog-content\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.610809 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss9dl\" (UniqueName: \"kubernetes.io/projected/2bfe5c0b-cde1-415e-846f-73947bfa62b4-kube-api-access-ss9dl\") pod \"community-operators-mfjrj\" (UID: \"2bfe5c0b-cde1-415e-846f-73947bfa62b4\") " pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.695960 4750 generic.go:334] "Generic (PLEG): container finished" podID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerID="84e7289668ac5c08c2fc2613e3fbcd8348cd606db6d0f5b844659e317311cb09" exitCode=0 Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.696024 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pf4pd" event={"ID":"a8a980d8-c735-4920-a9e3-ee6eb98b00b1","Type":"ContainerDied","Data":"84e7289668ac5c08c2fc2613e3fbcd8348cd606db6d0f5b844659e317311cb09"} Feb 20 00:13:58 crc kubenswrapper[4750]: I0220 00:13:58.709608 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6jfb" event={"ID":"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd","Type":"ContainerStarted","Data":"6f32bb7c2e7f7503aa81a1c536213a4be6b7c1b5a2ff3c38469d5204e52d2311"} Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:58.749743 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:58.945087 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5px4g"] Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.723198 4750 generic.go:334] "Generic (PLEG): container finished" podID="abac413a-7c83-4f73-8c5b-b1c0d50d5ebd" containerID="6f32bb7c2e7f7503aa81a1c536213a4be6b7c1b5a2ff3c38469d5204e52d2311" exitCode=0 Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.723537 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6jfb" event={"ID":"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd","Type":"ContainerDied","Data":"6f32bb7c2e7f7503aa81a1c536213a4be6b7c1b5a2ff3c38469d5204e52d2311"} Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.729694 4750 generic.go:334] "Generic (PLEG): container finished" podID="44fc4f50-400a-417e-9538-ef3dd5a66038" containerID="20a4ce7c5eaeae7fcc879f9d08fa06a7919275dcb1d70fe22d07182e286b9f06" exitCode=0 Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.729804 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5px4g" event={"ID":"44fc4f50-400a-417e-9538-ef3dd5a66038","Type":"ContainerDied","Data":"20a4ce7c5eaeae7fcc879f9d08fa06a7919275dcb1d70fe22d07182e286b9f06"} Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.729839 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5px4g" event={"ID":"44fc4f50-400a-417e-9538-ef3dd5a66038","Type":"ContainerStarted","Data":"476025df049ada0c5cfc0bbe1b7cdb1d3876ae766695bf9eb667021856bb6649"} Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.737724 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pf4pd" event={"ID":"a8a980d8-c735-4920-a9e3-ee6eb98b00b1","Type":"ContainerStarted","Data":"fcaffaf887b0a111113bca54d74f736b947d57f61f40f79ae53729ac53914c6a"} Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.738888 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mfjrj"] Feb 20 00:13:59 crc kubenswrapper[4750]: I0220 00:13:59.773659 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pf4pd" podStartSLOduration=3.278725723 podStartE2EDuration="4.773644677s" podCreationTimestamp="2026-02-20 00:13:55 +0000 UTC" firstStartedPulling="2026-02-20 00:13:57.691648339 +0000 UTC m=+321.886484398" lastFinishedPulling="2026-02-20 00:13:59.186567303 +0000 UTC m=+323.381403352" observedRunningTime="2026-02-20 00:13:59.773504323 +0000 UTC m=+323.968340372" watchObservedRunningTime="2026-02-20 00:13:59.773644677 +0000 UTC m=+323.968480726" Feb 20 00:13:59 crc kubenswrapper[4750]: W0220 00:13:59.811293 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bfe5c0b_cde1_415e_846f_73947bfa62b4.slice/crio-ba264383d48f476c0996df71e15981970cb49e55bb34534073e998e5efbb04ab WatchSource:0}: Error finding container ba264383d48f476c0996df71e15981970cb49e55bb34534073e998e5efbb04ab: Status 404 returned error can't find the container with id ba264383d48f476c0996df71e15981970cb49e55bb34534073e998e5efbb04ab Feb 20 00:14:00 crc kubenswrapper[4750]: I0220 00:14:00.744572 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5px4g" event={"ID":"44fc4f50-400a-417e-9538-ef3dd5a66038","Type":"ContainerStarted","Data":"dadebf109d42ef4f483c14fd55d4f62c3fab140083b818bc77c71edc415497f1"} Feb 20 00:14:00 crc kubenswrapper[4750]: I0220 00:14:00.746059 4750 generic.go:334] "Generic (PLEG): container finished" podID="2bfe5c0b-cde1-415e-846f-73947bfa62b4" containerID="7079b6617deecdf0c0e096de04de23786464c06d9304bb5fc7b4c2dc1b752316" exitCode=0 Feb 20 00:14:00 crc kubenswrapper[4750]: I0220 00:14:00.746136 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfjrj" event={"ID":"2bfe5c0b-cde1-415e-846f-73947bfa62b4","Type":"ContainerDied","Data":"7079b6617deecdf0c0e096de04de23786464c06d9304bb5fc7b4c2dc1b752316"} Feb 20 00:14:00 crc kubenswrapper[4750]: I0220 00:14:00.746164 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfjrj" event={"ID":"2bfe5c0b-cde1-415e-846f-73947bfa62b4","Type":"ContainerStarted","Data":"ba264383d48f476c0996df71e15981970cb49e55bb34534073e998e5efbb04ab"} Feb 20 00:14:00 crc kubenswrapper[4750]: I0220 00:14:00.748853 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m6jfb" event={"ID":"abac413a-7c83-4f73-8c5b-b1c0d50d5ebd","Type":"ContainerStarted","Data":"4aa2b430d297b35b0cda88f8daa2d1e0a7a3505b8bfd07fa620833c763aed169"} Feb 20 00:14:00 crc kubenswrapper[4750]: I0220 00:14:00.795625 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m6jfb" podStartSLOduration=3.2613347409999998 podStartE2EDuration="5.795605033s" podCreationTimestamp="2026-02-20 00:13:55 +0000 UTC" firstStartedPulling="2026-02-20 00:13:57.688878373 +0000 UTC m=+321.883714462" lastFinishedPulling="2026-02-20 00:14:00.223148705 +0000 UTC m=+324.417984754" observedRunningTime="2026-02-20 00:14:00.791804597 +0000 UTC m=+324.986640646" watchObservedRunningTime="2026-02-20 00:14:00.795605033 +0000 UTC m=+324.990441082" Feb 20 00:14:01 crc kubenswrapper[4750]: I0220 00:14:01.755898 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfjrj" event={"ID":"2bfe5c0b-cde1-415e-846f-73947bfa62b4","Type":"ContainerStarted","Data":"dab0562fe62af5ed7409d7a668aff60ee66aba0dfa966049902bdf5670e1e3d0"} Feb 20 00:14:01 crc kubenswrapper[4750]: I0220 00:14:01.758255 4750 generic.go:334] "Generic (PLEG): container finished" podID="44fc4f50-400a-417e-9538-ef3dd5a66038" containerID="dadebf109d42ef4f483c14fd55d4f62c3fab140083b818bc77c71edc415497f1" exitCode=0 Feb 20 00:14:01 crc kubenswrapper[4750]: I0220 00:14:01.758333 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5px4g" event={"ID":"44fc4f50-400a-417e-9538-ef3dd5a66038","Type":"ContainerDied","Data":"dadebf109d42ef4f483c14fd55d4f62c3fab140083b818bc77c71edc415497f1"} Feb 20 00:14:02 crc kubenswrapper[4750]: I0220 00:14:02.765929 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5px4g" event={"ID":"44fc4f50-400a-417e-9538-ef3dd5a66038","Type":"ContainerStarted","Data":"0452fbe0d6195ca772fbc900081589c1f89362f87472ebdae9bdfacb95283d5b"} Feb 20 00:14:02 crc kubenswrapper[4750]: I0220 00:14:02.768023 4750 generic.go:334] "Generic (PLEG): container finished" podID="2bfe5c0b-cde1-415e-846f-73947bfa62b4" containerID="dab0562fe62af5ed7409d7a668aff60ee66aba0dfa966049902bdf5670e1e3d0" exitCode=0 Feb 20 00:14:02 crc kubenswrapper[4750]: I0220 00:14:02.768104 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfjrj" event={"ID":"2bfe5c0b-cde1-415e-846f-73947bfa62b4","Type":"ContainerDied","Data":"dab0562fe62af5ed7409d7a668aff60ee66aba0dfa966049902bdf5670e1e3d0"} Feb 20 00:14:02 crc kubenswrapper[4750]: I0220 00:14:02.791907 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5px4g" podStartSLOduration=2.363689994 podStartE2EDuration="4.791884932s" podCreationTimestamp="2026-02-20 00:13:58 +0000 UTC" firstStartedPulling="2026-02-20 00:13:59.73159496 +0000 UTC m=+323.926431029" lastFinishedPulling="2026-02-20 00:14:02.159789888 +0000 UTC m=+326.354625967" observedRunningTime="2026-02-20 00:14:02.784924398 +0000 UTC m=+326.979760487" watchObservedRunningTime="2026-02-20 00:14:02.791884932 +0000 UTC m=+326.986720981" Feb 20 00:14:03 crc kubenswrapper[4750]: I0220 00:14:03.775416 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mfjrj" event={"ID":"2bfe5c0b-cde1-415e-846f-73947bfa62b4","Type":"ContainerStarted","Data":"0d9a089dd4720558b508f93840e64ec70e376ab4ac90e6624565b84c58f9f8db"} Feb 20 00:14:03 crc kubenswrapper[4750]: I0220 00:14:03.793046 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mfjrj" podStartSLOduration=3.146216625 podStartE2EDuration="5.793029709s" podCreationTimestamp="2026-02-20 00:13:58 +0000 UTC" firstStartedPulling="2026-02-20 00:14:00.74799519 +0000 UTC m=+324.942831239" lastFinishedPulling="2026-02-20 00:14:03.394808274 +0000 UTC m=+327.589644323" observedRunningTime="2026-02-20 00:14:03.790624572 +0000 UTC m=+327.985460631" watchObservedRunningTime="2026-02-20 00:14:03.793029709 +0000 UTC m=+327.987865758" Feb 20 00:14:06 crc kubenswrapper[4750]: I0220 00:14:06.136065 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:14:06 crc kubenswrapper[4750]: I0220 00:14:06.136158 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:14:06 crc kubenswrapper[4750]: I0220 00:14:06.324578 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:14:06 crc kubenswrapper[4750]: I0220 00:14:06.324645 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:14:06 crc kubenswrapper[4750]: I0220 00:14:06.364342 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:14:06 crc kubenswrapper[4750]: I0220 00:14:06.846990 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:14:07 crc kubenswrapper[4750]: I0220 00:14:07.179833 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-m6jfb" podUID="abac413a-7c83-4f73-8c5b-b1c0d50d5ebd" containerName="registry-server" probeResult="failure" output=< Feb 20 00:14:07 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Feb 20 00:14:07 crc kubenswrapper[4750]: > Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.545734 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.545947 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.618399 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.751029 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.751160 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.829331 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.889668 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5px4g" Feb 20 00:14:08 crc kubenswrapper[4750]: I0220 00:14:08.898993 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mfjrj" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.018021 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rmkb8"] Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.019212 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.032904 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rmkb8"] Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127009 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6ebae670-8ad1-47c2-aecf-4e59c81b0976-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127083 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwckw\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-kube-api-access-vwckw\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127139 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127164 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6ebae670-8ad1-47c2-aecf-4e59c81b0976-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127196 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-registry-tls\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127218 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-bound-sa-token\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127245 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ebae670-8ad1-47c2-aecf-4e59c81b0976-trusted-ca\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.127271 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6ebae670-8ad1-47c2-aecf-4e59c81b0976-registry-certificates\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.176405 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.191312 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.228188 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwckw\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-kube-api-access-vwckw\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.228252 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6ebae670-8ad1-47c2-aecf-4e59c81b0976-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.228309 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-registry-tls\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.228327 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-bound-sa-token\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.228347 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ebae670-8ad1-47c2-aecf-4e59c81b0976-trusted-ca\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.228362 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6ebae670-8ad1-47c2-aecf-4e59c81b0976-registry-certificates\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.228401 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6ebae670-8ad1-47c2-aecf-4e59c81b0976-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.229038 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6ebae670-8ad1-47c2-aecf-4e59c81b0976-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.229695 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ebae670-8ad1-47c2-aecf-4e59c81b0976-trusted-ca\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.229910 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6ebae670-8ad1-47c2-aecf-4e59c81b0976-registry-certificates\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.234718 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6ebae670-8ad1-47c2-aecf-4e59c81b0976-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.236500 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-registry-tls\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.251470 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwckw\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-kube-api-access-vwckw\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.255415 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m6jfb" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.255585 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6ebae670-8ad1-47c2-aecf-4e59c81b0976-bound-sa-token\") pod \"image-registry-66df7c8f76-rmkb8\" (UID: \"6ebae670-8ad1-47c2-aecf-4e59c81b0976\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.333620 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.743537 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rmkb8"] Feb 20 00:14:16 crc kubenswrapper[4750]: W0220 00:14:16.752880 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ebae670_8ad1_47c2_aecf_4e59c81b0976.slice/crio-819907f43a9266558c8b367ecbd51b1414eb719da7b42ab5bce110c992234de5 WatchSource:0}: Error finding container 819907f43a9266558c8b367ecbd51b1414eb719da7b42ab5bce110c992234de5: Status 404 returned error can't find the container with id 819907f43a9266558c8b367ecbd51b1414eb719da7b42ab5bce110c992234de5 Feb 20 00:14:16 crc kubenswrapper[4750]: I0220 00:14:16.848744 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" event={"ID":"6ebae670-8ad1-47c2-aecf-4e59c81b0976","Type":"ContainerStarted","Data":"819907f43a9266558c8b367ecbd51b1414eb719da7b42ab5bce110c992234de5"} Feb 20 00:14:17 crc kubenswrapper[4750]: I0220 00:14:17.854987 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" event={"ID":"6ebae670-8ad1-47c2-aecf-4e59c81b0976","Type":"ContainerStarted","Data":"7bf5c817dd627ee44f6156cecc7a51c63137d1b314615a8261cacf1ed38c97f1"} Feb 20 00:14:17 crc kubenswrapper[4750]: I0220 00:14:17.855152 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:17 crc kubenswrapper[4750]: I0220 00:14:17.881184 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" podStartSLOduration=1.881135972 podStartE2EDuration="1.881135972s" podCreationTimestamp="2026-02-20 00:14:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:14:17.879218318 +0000 UTC m=+342.074054377" watchObservedRunningTime="2026-02-20 00:14:17.881135972 +0000 UTC m=+342.075972021" Feb 20 00:14:26 crc kubenswrapper[4750]: I0220 00:14:26.669914 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:14:26 crc kubenswrapper[4750]: I0220 00:14:26.670494 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:14:36 crc kubenswrapper[4750]: I0220 00:14:36.346095 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-rmkb8" Feb 20 00:14:36 crc kubenswrapper[4750]: I0220 00:14:36.426301 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5r77d"] Feb 20 00:14:56 crc kubenswrapper[4750]: I0220 00:14:56.670332 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:14:56 crc kubenswrapper[4750]: I0220 00:14:56.670900 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.215646 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx"] Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.219327 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.223565 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.224248 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.227157 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx"] Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.321076 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67526005-e4a7-442d-ae55-603710921f68-secret-volume\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.321230 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67526005-e4a7-442d-ae55-603710921f68-config-volume\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.321569 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s75g7\" (UniqueName: \"kubernetes.io/projected/67526005-e4a7-442d-ae55-603710921f68-kube-api-access-s75g7\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.422735 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s75g7\" (UniqueName: \"kubernetes.io/projected/67526005-e4a7-442d-ae55-603710921f68-kube-api-access-s75g7\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.422871 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67526005-e4a7-442d-ae55-603710921f68-secret-volume\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.423021 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67526005-e4a7-442d-ae55-603710921f68-config-volume\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.425232 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67526005-e4a7-442d-ae55-603710921f68-config-volume\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.432380 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67526005-e4a7-442d-ae55-603710921f68-secret-volume\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.455373 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s75g7\" (UniqueName: \"kubernetes.io/projected/67526005-e4a7-442d-ae55-603710921f68-kube-api-access-s75g7\") pod \"collect-profiles-29525775-48jkx\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.541887 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:00 crc kubenswrapper[4750]: I0220 00:15:00.762523 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx"] Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.162632 4750 generic.go:334] "Generic (PLEG): container finished" podID="67526005-e4a7-442d-ae55-603710921f68" containerID="06d5a6dd52c38239d48e0e3b0a627edd343198542eac98cc984350a36fd2f285" exitCode=0 Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.162680 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" event={"ID":"67526005-e4a7-442d-ae55-603710921f68","Type":"ContainerDied","Data":"06d5a6dd52c38239d48e0e3b0a627edd343198542eac98cc984350a36fd2f285"} Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.162966 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" event={"ID":"67526005-e4a7-442d-ae55-603710921f68","Type":"ContainerStarted","Data":"deb75eda77119c635d0994444b354d0de5f5c0353a268ef9167db3734b049365"} Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.477436 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" podUID="5d7c67d4-013f-494c-88dc-079a3accf05b" containerName="registry" containerID="cri-o://7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334" gracePeriod=30 Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.871942 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954058 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xkkr\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-kube-api-access-9xkkr\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954104 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-tls\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954152 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-certificates\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954316 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954366 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-trusted-ca\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954473 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5d7c67d4-013f-494c-88dc-079a3accf05b-installation-pull-secrets\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954514 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5d7c67d4-013f-494c-88dc-079a3accf05b-ca-trust-extracted\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.954551 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-bound-sa-token\") pod \"5d7c67d4-013f-494c-88dc-079a3accf05b\" (UID: \"5d7c67d4-013f-494c-88dc-079a3accf05b\") " Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.955246 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.956322 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.961856 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7c67d4-013f-494c-88dc-079a3accf05b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.962431 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-kube-api-access-9xkkr" (OuterVolumeSpecName: "kube-api-access-9xkkr") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "kube-api-access-9xkkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.964942 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.964994 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.965796 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 00:15:01 crc kubenswrapper[4750]: I0220 00:15:01.978776 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d7c67d4-013f-494c-88dc-079a3accf05b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "5d7c67d4-013f-494c-88dc-079a3accf05b" (UID: "5d7c67d4-013f-494c-88dc-079a3accf05b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.056444 4750 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.056506 4750 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5d7c67d4-013f-494c-88dc-079a3accf05b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.056531 4750 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5d7c67d4-013f-494c-88dc-079a3accf05b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.056562 4750 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.056579 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xkkr\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-kube-api-access-9xkkr\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.056595 4750 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.056613 4750 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5d7c67d4-013f-494c-88dc-079a3accf05b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.178435 4750 generic.go:334] "Generic (PLEG): container finished" podID="5d7c67d4-013f-494c-88dc-079a3accf05b" containerID="7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334" exitCode=0 Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.178508 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" event={"ID":"5d7c67d4-013f-494c-88dc-079a3accf05b","Type":"ContainerDied","Data":"7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334"} Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.178545 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.178579 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5r77d" event={"ID":"5d7c67d4-013f-494c-88dc-079a3accf05b","Type":"ContainerDied","Data":"d3596ec8045423b04fc5289d0279b810bab12bc707ac4fbedc609ee91a94c1c5"} Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.178613 4750 scope.go:117] "RemoveContainer" containerID="7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.207927 4750 scope.go:117] "RemoveContainer" containerID="7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334" Feb 20 00:15:02 crc kubenswrapper[4750]: E0220 00:15:02.208659 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334\": container with ID starting with 7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334 not found: ID does not exist" containerID="7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.208699 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334"} err="failed to get container status \"7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334\": rpc error: code = NotFound desc = could not find container \"7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334\": container with ID starting with 7918b4e6f16eefdd5328275976e29e50bd85de2c62387ea66015c55b5deac334 not found: ID does not exist" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.231318 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5r77d"] Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.235159 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5r77d"] Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.404460 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.461289 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67526005-e4a7-442d-ae55-603710921f68-secret-volume\") pod \"67526005-e4a7-442d-ae55-603710921f68\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.461378 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s75g7\" (UniqueName: \"kubernetes.io/projected/67526005-e4a7-442d-ae55-603710921f68-kube-api-access-s75g7\") pod \"67526005-e4a7-442d-ae55-603710921f68\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.461495 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67526005-e4a7-442d-ae55-603710921f68-config-volume\") pod \"67526005-e4a7-442d-ae55-603710921f68\" (UID: \"67526005-e4a7-442d-ae55-603710921f68\") " Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.462352 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67526005-e4a7-442d-ae55-603710921f68-config-volume" (OuterVolumeSpecName: "config-volume") pod "67526005-e4a7-442d-ae55-603710921f68" (UID: "67526005-e4a7-442d-ae55-603710921f68"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.465883 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67526005-e4a7-442d-ae55-603710921f68-kube-api-access-s75g7" (OuterVolumeSpecName: "kube-api-access-s75g7") pod "67526005-e4a7-442d-ae55-603710921f68" (UID: "67526005-e4a7-442d-ae55-603710921f68"). InnerVolumeSpecName "kube-api-access-s75g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.466051 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67526005-e4a7-442d-ae55-603710921f68-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "67526005-e4a7-442d-ae55-603710921f68" (UID: "67526005-e4a7-442d-ae55-603710921f68"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.562330 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s75g7\" (UniqueName: \"kubernetes.io/projected/67526005-e4a7-442d-ae55-603710921f68-kube-api-access-s75g7\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.562364 4750 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67526005-e4a7-442d-ae55-603710921f68-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.562372 4750 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67526005-e4a7-442d-ae55-603710921f68-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 00:15:02 crc kubenswrapper[4750]: I0220 00:15:02.566460 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d7c67d4-013f-494c-88dc-079a3accf05b" path="/var/lib/kubelet/pods/5d7c67d4-013f-494c-88dc-079a3accf05b/volumes" Feb 20 00:15:03 crc kubenswrapper[4750]: I0220 00:15:03.188486 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" Feb 20 00:15:03 crc kubenswrapper[4750]: I0220 00:15:03.188496 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29525775-48jkx" event={"ID":"67526005-e4a7-442d-ae55-603710921f68","Type":"ContainerDied","Data":"deb75eda77119c635d0994444b354d0de5f5c0353a268ef9167db3734b049365"} Feb 20 00:15:03 crc kubenswrapper[4750]: I0220 00:15:03.189207 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="deb75eda77119c635d0994444b354d0de5f5c0353a268ef9167db3734b049365" Feb 20 00:15:26 crc kubenswrapper[4750]: I0220 00:15:26.669817 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:15:26 crc kubenswrapper[4750]: I0220 00:15:26.670563 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:15:26 crc kubenswrapper[4750]: I0220 00:15:26.670618 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:15:26 crc kubenswrapper[4750]: I0220 00:15:26.671264 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f3a4507cbe00af55a5fb388980bbf31c52a319e48434ce9fcf1faccc79339c1"} pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 00:15:26 crc kubenswrapper[4750]: I0220 00:15:26.671345 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" containerID="cri-o://7f3a4507cbe00af55a5fb388980bbf31c52a319e48434ce9fcf1faccc79339c1" gracePeriod=600 Feb 20 00:15:27 crc kubenswrapper[4750]: I0220 00:15:27.393087 4750 generic.go:334] "Generic (PLEG): container finished" podID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerID="7f3a4507cbe00af55a5fb388980bbf31c52a319e48434ce9fcf1faccc79339c1" exitCode=0 Feb 20 00:15:27 crc kubenswrapper[4750]: I0220 00:15:27.393186 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerDied","Data":"7f3a4507cbe00af55a5fb388980bbf31c52a319e48434ce9fcf1faccc79339c1"} Feb 20 00:15:27 crc kubenswrapper[4750]: I0220 00:15:27.393770 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"4ba5fe5180b1882694da5fe6cd20529927f075cad71b329d6a7948bdcfb58cc2"} Feb 20 00:15:27 crc kubenswrapper[4750]: I0220 00:15:27.393802 4750 scope.go:117] "RemoveContainer" containerID="67ec509472a6c3018f932c1fbc9cd567cb77627ea31c909dc128429058436952" Feb 20 00:17:26 crc kubenswrapper[4750]: I0220 00:17:26.669712 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:17:26 crc kubenswrapper[4750]: I0220 00:17:26.670428 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.422227 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hkqgt"] Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.423754 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-controller" containerID="cri-o://289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.424154 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="northd" containerID="cri-o://28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.424465 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="sbdb" containerID="cri-o://694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.424534 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="nbdb" containerID="cri-o://ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.424571 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.424686 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-acl-logging" containerID="cri-o://83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.424788 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-node" containerID="cri-o://192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.519408 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" containerID="cri-o://411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" gracePeriod=30 Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.826865 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/3.log" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.829756 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovn-acl-logging/0.log" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.830302 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovn-controller/0.log" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.831062 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.864928 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-ovn-kubernetes\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865037 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-log-socket\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865072 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-var-lib-openvswitch\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865119 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovn-node-metrics-cert\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865177 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-slash\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865208 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-systemd\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865233 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-systemd-units\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865294 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-node-log\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865336 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-netns\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865362 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865390 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-bin\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865436 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-config\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865470 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm5lm\" (UniqueName: \"kubernetes.io/projected/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-kube-api-access-xm5lm\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865497 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-script-lib\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865540 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-openvswitch\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865570 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-etc-openvswitch\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865593 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-ovn\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865628 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-env-overrides\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865653 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-netd\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.865683 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-kubelet\") pod \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\" (UID: \"17ae630a-7ade-407c-b93c-e4bae6d0f8e3\") " Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866105 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866190 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866222 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-log-socket" (OuterVolumeSpecName: "log-socket") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866247 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866512 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-node-log" (OuterVolumeSpecName: "node-log") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866599 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-slash" (OuterVolumeSpecName: "host-slash") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866610 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866685 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866697 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866736 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866816 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866821 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866848 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.866975 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.867042 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.867100 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.867144 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.875641 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-kube-api-access-xm5lm" (OuterVolumeSpecName: "kube-api-access-xm5lm") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "kube-api-access-xm5lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.876749 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.887399 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "17ae630a-7ade-407c-b93c-e4bae6d0f8e3" (UID: "17ae630a-7ade-407c-b93c-e4bae6d0f8e3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908309 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-n7f2d"] Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908665 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67526005-e4a7-442d-ae55-603710921f68" containerName="collect-profiles" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908695 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="67526005-e4a7-442d-ae55-603710921f68" containerName="collect-profiles" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908712 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908722 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908733 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908741 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908752 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kubecfg-setup" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908760 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kubecfg-setup" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908770 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908777 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908789 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908796 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908808 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="northd" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908815 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="northd" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908826 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7c67d4-013f-494c-88dc-079a3accf05b" containerName="registry" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908833 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7c67d4-013f-494c-88dc-079a3accf05b" containerName="registry" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908840 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908847 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908858 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908866 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908874 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="nbdb" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908880 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="nbdb" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908888 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-node" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908895 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-node" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908903 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="sbdb" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908909 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="sbdb" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.908918 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-acl-logging" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.908924 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-acl-logging" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909026 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="67526005-e4a7-442d-ae55-603710921f68" containerName="collect-profiles" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909041 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909050 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="sbdb" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909058 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909065 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909071 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909078 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovn-acl-logging" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909090 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-node" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909099 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909106 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="northd" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909117 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="nbdb" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909144 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909153 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7c67d4-013f-494c-88dc-079a3accf05b" containerName="registry" Feb 20 00:17:50 crc kubenswrapper[4750]: E0220 00:17:50.909260 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909268 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.909368 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerName="ovnkube-controller" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.911730 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.966869 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-ovn\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.966925 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.966953 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovnkube-config\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.966980 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-kubelet\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967005 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-cni-netd\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967049 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovn-node-metrics-cert\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967237 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-node-log\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967363 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-systemd-units\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967421 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-env-overrides\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967512 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-run-netns\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967546 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-etc-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967569 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-var-lib-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967741 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-systemd\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967865 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-log-socket\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967908 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967937 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7dfx\" (UniqueName: \"kubernetes.io/projected/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-kube-api-access-g7dfx\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.967969 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-cni-bin\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968058 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovnkube-script-lib\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968177 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-run-ovn-kubernetes\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968249 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-slash\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968381 4750 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968407 4750 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968428 4750 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-node-log\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968448 4750 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968466 4750 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968488 4750 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968508 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968528 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm5lm\" (UniqueName: \"kubernetes.io/projected/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-kube-api-access-xm5lm\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968550 4750 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968569 4750 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968588 4750 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968612 4750 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968630 4750 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968648 4750 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968668 4750 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968690 4750 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968708 4750 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-log-socket\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968726 4750 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968746 4750 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:50 crc kubenswrapper[4750]: I0220 00:17:50.968764 4750 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/17ae630a-7ade-407c-b93c-e4bae6d0f8e3-host-slash\") on node \"crc\" DevicePath \"\"" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070256 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovnkube-config\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070374 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-kubelet\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070439 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-cni-netd\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070483 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovn-node-metrics-cert\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070533 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-node-log\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070558 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-kubelet\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070603 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-systemd-units\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070641 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-cni-netd\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070685 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-systemd-units\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070651 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-env-overrides\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070652 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-node-log\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070766 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-run-netns\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070803 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-etc-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070839 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-var-lib-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070893 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-systemd\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070935 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-log-socket\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070958 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovnkube-config\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.070971 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071018 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-var-lib-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071017 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7dfx\" (UniqueName: \"kubernetes.io/projected/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-kube-api-access-g7dfx\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071065 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-cni-bin\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071085 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovnkube-script-lib\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071144 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-run-ovn-kubernetes\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071170 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-slash\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071197 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-ovn\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071215 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071293 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071318 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-run-netns\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071342 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-run-ovn-kubernetes\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071385 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-etc-openvswitch\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071398 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-slash\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071447 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-cni-bin\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071452 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071444 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-log-socket\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071463 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-ovn\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071462 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-run-systemd\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.071705 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-env-overrides\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.072854 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovnkube-script-lib\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.075566 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-ovn-node-metrics-cert\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.097856 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7dfx\" (UniqueName: \"kubernetes.io/projected/ceb9c632-168e-47b9-b85e-75ef1fcfbfda-kube-api-access-g7dfx\") pod \"ovnkube-node-n7f2d\" (UID: \"ceb9c632-168e-47b9-b85e-75ef1fcfbfda\") " pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.226032 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.376336 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovnkube-controller/3.log" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.380761 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovn-acl-logging/0.log" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.381787 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hkqgt_17ae630a-7ade-407c-b93c-e4bae6d0f8e3/ovn-controller/0.log" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382581 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" exitCode=0 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382623 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" exitCode=0 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382644 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" exitCode=0 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382663 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" exitCode=0 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382680 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" exitCode=0 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382700 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" exitCode=0 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382717 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" exitCode=143 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382731 4750 generic.go:334] "Generic (PLEG): container finished" podID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" containerID="289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" exitCode=143 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382807 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382866 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382896 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382922 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382952 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.382977 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383001 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383021 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383033 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383044 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383097 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383109 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383156 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383173 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383188 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383209 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383233 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383249 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383263 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383277 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383296 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383309 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383325 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383341 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383355 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383369 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383389 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383412 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383431 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383447 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383462 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383476 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383490 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383494 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383563 4750 scope.go:117] "RemoveContainer" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383504 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383665 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383682 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383697 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383718 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hkqgt" event={"ID":"17ae630a-7ade-407c-b93c-e4bae6d0f8e3","Type":"ContainerDied","Data":"22ef73bc6fd4ef403a25b7711e8495189441418f3c270591532b3048b06e9e63"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383742 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383760 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383774 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383788 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383802 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383814 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383826 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383838 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383853 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.383864 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.388356 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/2.log" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.394088 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/1.log" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.394206 4750 generic.go:334] "Generic (PLEG): container finished" podID="1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030" containerID="2edf2c4409f090699cacb0b51a9257cd96eccff1387502cefc49f5308693f05d" exitCode=2 Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.394326 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerDied","Data":"2edf2c4409f090699cacb0b51a9257cd96eccff1387502cefc49f5308693f05d"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.394375 4750 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.395314 4750 scope.go:117] "RemoveContainer" containerID="2edf2c4409f090699cacb0b51a9257cd96eccff1387502cefc49f5308693f05d" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.395696 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dm8lb_openshift-multus(1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030)\"" pod="openshift-multus/multus-dm8lb" podUID="1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.413897 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"53e8d52f635dcd2303c445effdd7897535c2ac5c991e98c68c4ec16c6d73ccb9"} Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.435954 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.470519 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hkqgt"] Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.477712 4750 scope.go:117] "RemoveContainer" containerID="694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.480273 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hkqgt"] Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.535542 4750 scope.go:117] "RemoveContainer" containerID="ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.552682 4750 scope.go:117] "RemoveContainer" containerID="28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.580146 4750 scope.go:117] "RemoveContainer" containerID="540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.621586 4750 scope.go:117] "RemoveContainer" containerID="192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.644688 4750 scope.go:117] "RemoveContainer" containerID="83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.664964 4750 scope.go:117] "RemoveContainer" containerID="289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.680609 4750 scope.go:117] "RemoveContainer" containerID="9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.702907 4750 scope.go:117] "RemoveContainer" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.703476 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": container with ID starting with 411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2 not found: ID does not exist" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.703548 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} err="failed to get container status \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": rpc error: code = NotFound desc = could not find container \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": container with ID starting with 411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.703592 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.704195 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": container with ID starting with 44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5 not found: ID does not exist" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.704327 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} err="failed to get container status \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": rpc error: code = NotFound desc = could not find container \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": container with ID starting with 44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.704434 4750 scope.go:117] "RemoveContainer" containerID="694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.705107 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": container with ID starting with 694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31 not found: ID does not exist" containerID="694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.705270 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} err="failed to get container status \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": rpc error: code = NotFound desc = could not find container \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": container with ID starting with 694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.705388 4750 scope.go:117] "RemoveContainer" containerID="ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.705990 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": container with ID starting with ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2 not found: ID does not exist" containerID="ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.706064 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} err="failed to get container status \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": rpc error: code = NotFound desc = could not find container \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": container with ID starting with ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.706100 4750 scope.go:117] "RemoveContainer" containerID="28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.706495 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": container with ID starting with 28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292 not found: ID does not exist" containerID="28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.706544 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} err="failed to get container status \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": rpc error: code = NotFound desc = could not find container \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": container with ID starting with 28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.706575 4750 scope.go:117] "RemoveContainer" containerID="540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.707108 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": container with ID starting with 540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95 not found: ID does not exist" containerID="540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.707185 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} err="failed to get container status \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": rpc error: code = NotFound desc = could not find container \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": container with ID starting with 540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.707211 4750 scope.go:117] "RemoveContainer" containerID="192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.707622 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": container with ID starting with 192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40 not found: ID does not exist" containerID="192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.707736 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} err="failed to get container status \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": rpc error: code = NotFound desc = could not find container \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": container with ID starting with 192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.707826 4750 scope.go:117] "RemoveContainer" containerID="83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.708531 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": container with ID starting with 83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b not found: ID does not exist" containerID="83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.708573 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} err="failed to get container status \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": rpc error: code = NotFound desc = could not find container \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": container with ID starting with 83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.708600 4750 scope.go:117] "RemoveContainer" containerID="289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.709153 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": container with ID starting with 289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1 not found: ID does not exist" containerID="289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.709260 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} err="failed to get container status \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": rpc error: code = NotFound desc = could not find container \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": container with ID starting with 289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.709349 4750 scope.go:117] "RemoveContainer" containerID="9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5" Feb 20 00:17:51 crc kubenswrapper[4750]: E0220 00:17:51.710551 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": container with ID starting with 9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5 not found: ID does not exist" containerID="9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.710599 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} err="failed to get container status \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": rpc error: code = NotFound desc = could not find container \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": container with ID starting with 9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.710629 4750 scope.go:117] "RemoveContainer" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.711115 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} err="failed to get container status \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": rpc error: code = NotFound desc = could not find container \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": container with ID starting with 411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.711239 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.711794 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} err="failed to get container status \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": rpc error: code = NotFound desc = could not find container \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": container with ID starting with 44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.711850 4750 scope.go:117] "RemoveContainer" containerID="694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.712403 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} err="failed to get container status \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": rpc error: code = NotFound desc = could not find container \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": container with ID starting with 694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.712565 4750 scope.go:117] "RemoveContainer" containerID="ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.713185 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} err="failed to get container status \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": rpc error: code = NotFound desc = could not find container \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": container with ID starting with ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.713228 4750 scope.go:117] "RemoveContainer" containerID="28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.713654 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} err="failed to get container status \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": rpc error: code = NotFound desc = could not find container \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": container with ID starting with 28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.713698 4750 scope.go:117] "RemoveContainer" containerID="540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.714254 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} err="failed to get container status \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": rpc error: code = NotFound desc = could not find container \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": container with ID starting with 540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.714324 4750 scope.go:117] "RemoveContainer" containerID="192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.714713 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} err="failed to get container status \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": rpc error: code = NotFound desc = could not find container \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": container with ID starting with 192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.714759 4750 scope.go:117] "RemoveContainer" containerID="83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.715289 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} err="failed to get container status \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": rpc error: code = NotFound desc = could not find container \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": container with ID starting with 83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.715393 4750 scope.go:117] "RemoveContainer" containerID="289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.715839 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} err="failed to get container status \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": rpc error: code = NotFound desc = could not find container \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": container with ID starting with 289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.715948 4750 scope.go:117] "RemoveContainer" containerID="9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.716653 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} err="failed to get container status \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": rpc error: code = NotFound desc = could not find container \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": container with ID starting with 9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.716692 4750 scope.go:117] "RemoveContainer" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.717252 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} err="failed to get container status \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": rpc error: code = NotFound desc = could not find container \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": container with ID starting with 411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.717287 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.717920 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} err="failed to get container status \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": rpc error: code = NotFound desc = could not find container \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": container with ID starting with 44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.717967 4750 scope.go:117] "RemoveContainer" containerID="694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.718506 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} err="failed to get container status \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": rpc error: code = NotFound desc = could not find container \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": container with ID starting with 694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.718588 4750 scope.go:117] "RemoveContainer" containerID="ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.719081 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} err="failed to get container status \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": rpc error: code = NotFound desc = could not find container \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": container with ID starting with ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.719154 4750 scope.go:117] "RemoveContainer" containerID="28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.719485 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} err="failed to get container status \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": rpc error: code = NotFound desc = could not find container \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": container with ID starting with 28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.719516 4750 scope.go:117] "RemoveContainer" containerID="540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.719906 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} err="failed to get container status \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": rpc error: code = NotFound desc = could not find container \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": container with ID starting with 540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.719945 4750 scope.go:117] "RemoveContainer" containerID="192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.720378 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} err="failed to get container status \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": rpc error: code = NotFound desc = could not find container \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": container with ID starting with 192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.720406 4750 scope.go:117] "RemoveContainer" containerID="83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.720856 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} err="failed to get container status \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": rpc error: code = NotFound desc = could not find container \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": container with ID starting with 83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.720899 4750 scope.go:117] "RemoveContainer" containerID="289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.721343 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} err="failed to get container status \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": rpc error: code = NotFound desc = could not find container \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": container with ID starting with 289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.721380 4750 scope.go:117] "RemoveContainer" containerID="9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.721783 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} err="failed to get container status \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": rpc error: code = NotFound desc = could not find container \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": container with ID starting with 9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.721874 4750 scope.go:117] "RemoveContainer" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.722680 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} err="failed to get container status \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": rpc error: code = NotFound desc = could not find container \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": container with ID starting with 411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.722718 4750 scope.go:117] "RemoveContainer" containerID="44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.723088 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5"} err="failed to get container status \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": rpc error: code = NotFound desc = could not find container \"44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5\": container with ID starting with 44c340f072348c90136a7e94c3ac001cf26c5e722373c25400e4686fc4f1daa5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.723156 4750 scope.go:117] "RemoveContainer" containerID="694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.723747 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31"} err="failed to get container status \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": rpc error: code = NotFound desc = could not find container \"694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31\": container with ID starting with 694566f27233d1935cf5a5275c71a4eccabb01d85e97147cea4cdc6f97833e31 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.723780 4750 scope.go:117] "RemoveContainer" containerID="ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.724226 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2"} err="failed to get container status \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": rpc error: code = NotFound desc = could not find container \"ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2\": container with ID starting with ea0e59a0c9a58714abd0fa6422dc1ee5b1c7fb64f3e1bbcc17c1b5478bfc9be2 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.724265 4750 scope.go:117] "RemoveContainer" containerID="28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.724719 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292"} err="failed to get container status \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": rpc error: code = NotFound desc = could not find container \"28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292\": container with ID starting with 28510a4a8cac9fd5b99f52002a99b9228b458596b081f3bd0cafbdf5cbb29292 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.724752 4750 scope.go:117] "RemoveContainer" containerID="540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.725385 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95"} err="failed to get container status \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": rpc error: code = NotFound desc = could not find container \"540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95\": container with ID starting with 540e5506dc784adeb4e2a681a9c20e860a59d81632bd7197c538736294040d95 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.725429 4750 scope.go:117] "RemoveContainer" containerID="192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.725872 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40"} err="failed to get container status \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": rpc error: code = NotFound desc = could not find container \"192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40\": container with ID starting with 192667ca363982fd88b036f36c9c1f47438c063238ea217ce525e40898430a40 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.725913 4750 scope.go:117] "RemoveContainer" containerID="83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.727574 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b"} err="failed to get container status \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": rpc error: code = NotFound desc = could not find container \"83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b\": container with ID starting with 83a2db212e5e06d5e2a4875538562d271ccf5a3698a7a82bed0fc2ba5b954d3b not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.727623 4750 scope.go:117] "RemoveContainer" containerID="289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.728175 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1"} err="failed to get container status \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": rpc error: code = NotFound desc = could not find container \"289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1\": container with ID starting with 289f4404ce057768b2c0c524b9e7d95abffc761dd617a5e35a8f7748f3bb15b1 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.728214 4750 scope.go:117] "RemoveContainer" containerID="9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.728621 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5"} err="failed to get container status \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": rpc error: code = NotFound desc = could not find container \"9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5\": container with ID starting with 9cf06936e63b614b34924dac00937b114c1dba90b1acacc66436c002c902b4c5 not found: ID does not exist" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.728651 4750 scope.go:117] "RemoveContainer" containerID="411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2" Feb 20 00:17:51 crc kubenswrapper[4750]: I0220 00:17:51.729052 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2"} err="failed to get container status \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": rpc error: code = NotFound desc = could not find container \"411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2\": container with ID starting with 411fd9cbd99a5bc9c3609cbeea888743953017e7af241fd5a6591db687559cc2 not found: ID does not exist" Feb 20 00:17:52 crc kubenswrapper[4750]: I0220 00:17:52.421533 4750 generic.go:334] "Generic (PLEG): container finished" podID="ceb9c632-168e-47b9-b85e-75ef1fcfbfda" containerID="b324c7865f590d4209eb071ac3b1f704c38c7a2b938b6dbdb213bc70094604b9" exitCode=0 Feb 20 00:17:52 crc kubenswrapper[4750]: I0220 00:17:52.421643 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerDied","Data":"b324c7865f590d4209eb071ac3b1f704c38c7a2b938b6dbdb213bc70094604b9"} Feb 20 00:17:52 crc kubenswrapper[4750]: I0220 00:17:52.572506 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17ae630a-7ade-407c-b93c-e4bae6d0f8e3" path="/var/lib/kubelet/pods/17ae630a-7ade-407c-b93c-e4bae6d0f8e3/volumes" Feb 20 00:17:53 crc kubenswrapper[4750]: I0220 00:17:53.433074 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"1034bcac3dc4aca450f53df0c05b7dfd3b9f46694f36342e4084fba8b33014e2"} Feb 20 00:17:53 crc kubenswrapper[4750]: I0220 00:17:53.433581 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"cf76f441ec4ebbe51a7d3c6ced8645c0f8cf59b543ea2041006da1f6ea9e2653"} Feb 20 00:17:53 crc kubenswrapper[4750]: I0220 00:17:53.433607 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"03ba745dac0e4ba15a1c709dbf491540095a3a8e39f25303403e35ab588ad1b7"} Feb 20 00:17:53 crc kubenswrapper[4750]: I0220 00:17:53.433631 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"13e49ad1ce36756a6b0e913a266c5aa74af3f1aff55dcc9144e3bd3c907b83fb"} Feb 20 00:17:53 crc kubenswrapper[4750]: I0220 00:17:53.433649 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"8f8f7bab79022a2b1ffb6f348c0451e24bb5e5dcd15bae5a5f3ac217f6afd1b1"} Feb 20 00:17:53 crc kubenswrapper[4750]: I0220 00:17:53.433666 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"398e9cb887aecd07e352a424632af938219dd9f7e9f342e1db6a0c93c5f659cd"} Feb 20 00:17:56 crc kubenswrapper[4750]: I0220 00:17:56.459586 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"3e9ea7845a744b1ba92047e926ca2f41489dc440f570268441cd3af3f5db692a"} Feb 20 00:17:56 crc kubenswrapper[4750]: I0220 00:17:56.669955 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:17:56 crc kubenswrapper[4750]: I0220 00:17:56.670025 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:17:58 crc kubenswrapper[4750]: I0220 00:17:58.493461 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" event={"ID":"ceb9c632-168e-47b9-b85e-75ef1fcfbfda","Type":"ContainerStarted","Data":"61606ef167a4d70d12da5c94766202e5fef1e2973a981a5d4f53e8a7ee6e864d"} Feb 20 00:17:58 crc kubenswrapper[4750]: I0220 00:17:58.493693 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:58 crc kubenswrapper[4750]: I0220 00:17:58.493705 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:58 crc kubenswrapper[4750]: I0220 00:17:58.493714 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:58 crc kubenswrapper[4750]: I0220 00:17:58.521788 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:17:58 crc kubenswrapper[4750]: I0220 00:17:58.529880 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" podStartSLOduration=8.529853621 podStartE2EDuration="8.529853621s" podCreationTimestamp="2026-02-20 00:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:17:58.524676961 +0000 UTC m=+562.719513010" watchObservedRunningTime="2026-02-20 00:17:58.529853621 +0000 UTC m=+562.724689680" Feb 20 00:17:58 crc kubenswrapper[4750]: I0220 00:17:58.574153 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:18:05 crc kubenswrapper[4750]: I0220 00:18:05.561514 4750 scope.go:117] "RemoveContainer" containerID="2edf2c4409f090699cacb0b51a9257cd96eccff1387502cefc49f5308693f05d" Feb 20 00:18:05 crc kubenswrapper[4750]: E0220 00:18:05.562956 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dm8lb_openshift-multus(1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030)\"" pod="openshift-multus/multus-dm8lb" podUID="1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030" Feb 20 00:18:17 crc kubenswrapper[4750]: I0220 00:18:17.560414 4750 scope.go:117] "RemoveContainer" containerID="2edf2c4409f090699cacb0b51a9257cd96eccff1387502cefc49f5308693f05d" Feb 20 00:18:18 crc kubenswrapper[4750]: I0220 00:18:18.633798 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/2.log" Feb 20 00:18:18 crc kubenswrapper[4750]: I0220 00:18:18.634752 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/1.log" Feb 20 00:18:18 crc kubenswrapper[4750]: I0220 00:18:18.634851 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dm8lb" event={"ID":"1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030","Type":"ContainerStarted","Data":"b06908ac8324898f99695229ce41d907508e60a156189ff4a743263cc913902f"} Feb 20 00:18:21 crc kubenswrapper[4750]: I0220 00:18:21.269871 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-n7f2d" Feb 20 00:18:26 crc kubenswrapper[4750]: I0220 00:18:26.670249 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:18:26 crc kubenswrapper[4750]: I0220 00:18:26.670682 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:18:26 crc kubenswrapper[4750]: I0220 00:18:26.670749 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:18:26 crc kubenswrapper[4750]: I0220 00:18:26.671596 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ba5fe5180b1882694da5fe6cd20529927f075cad71b329d6a7948bdcfb58cc2"} pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 00:18:26 crc kubenswrapper[4750]: I0220 00:18:26.671701 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" containerID="cri-o://4ba5fe5180b1882694da5fe6cd20529927f075cad71b329d6a7948bdcfb58cc2" gracePeriod=600 Feb 20 00:18:27 crc kubenswrapper[4750]: I0220 00:18:27.703951 4750 generic.go:334] "Generic (PLEG): container finished" podID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerID="4ba5fe5180b1882694da5fe6cd20529927f075cad71b329d6a7948bdcfb58cc2" exitCode=0 Feb 20 00:18:27 crc kubenswrapper[4750]: I0220 00:18:27.704013 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerDied","Data":"4ba5fe5180b1882694da5fe6cd20529927f075cad71b329d6a7948bdcfb58cc2"} Feb 20 00:18:27 crc kubenswrapper[4750]: I0220 00:18:27.704495 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"2df5c0a1c7798ac15bd3ceb64e0a208a9488dd206f4f7114aee925599addd0eb"} Feb 20 00:18:27 crc kubenswrapper[4750]: I0220 00:18:27.704542 4750 scope.go:117] "RemoveContainer" containerID="7f3a4507cbe00af55a5fb388980bbf31c52a319e48434ce9fcf1faccc79339c1" Feb 20 00:18:36 crc kubenswrapper[4750]: I0220 00:18:36.847654 4750 scope.go:117] "RemoveContainer" containerID="8bac1e684c9212698e2953131f7ed28fe4bc174889a5c260f096b0016e2a7002" Feb 20 00:18:37 crc kubenswrapper[4750]: I0220 00:18:37.792444 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dm8lb_1acbd435-6fc9-40bf-b2b4-1f8ff1b9c030/kube-multus/2.log" Feb 20 00:18:50 crc kubenswrapper[4750]: I0220 00:18:50.467155 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pf4pd"] Feb 20 00:18:50 crc kubenswrapper[4750]: I0220 00:18:50.468431 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pf4pd" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="registry-server" containerID="cri-o://fcaffaf887b0a111113bca54d74f736b947d57f61f40f79ae53729ac53914c6a" gracePeriod=30 Feb 20 00:18:50 crc kubenswrapper[4750]: I0220 00:18:50.899022 4750 generic.go:334] "Generic (PLEG): container finished" podID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerID="fcaffaf887b0a111113bca54d74f736b947d57f61f40f79ae53729ac53914c6a" exitCode=0 Feb 20 00:18:50 crc kubenswrapper[4750]: I0220 00:18:50.899311 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pf4pd" event={"ID":"a8a980d8-c735-4920-a9e3-ee6eb98b00b1","Type":"ContainerDied","Data":"fcaffaf887b0a111113bca54d74f736b947d57f61f40f79ae53729ac53914c6a"} Feb 20 00:18:50 crc kubenswrapper[4750]: I0220 00:18:50.899629 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pf4pd" event={"ID":"a8a980d8-c735-4920-a9e3-ee6eb98b00b1","Type":"ContainerDied","Data":"390e82ed07e0638b7e54cd0b81d144096e50037cb2f9db3dc0999d51ad99c5f7"} Feb 20 00:18:50 crc kubenswrapper[4750]: I0220 00:18:50.899657 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="390e82ed07e0638b7e54cd0b81d144096e50037cb2f9db3dc0999d51ad99c5f7" Feb 20 00:18:50 crc kubenswrapper[4750]: I0220 00:18:50.924477 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.063255 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8fl5\" (UniqueName: \"kubernetes.io/projected/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-kube-api-access-g8fl5\") pod \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.063386 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-utilities\") pod \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.063518 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-catalog-content\") pod \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\" (UID: \"a8a980d8-c735-4920-a9e3-ee6eb98b00b1\") " Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.064760 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-utilities" (OuterVolumeSpecName: "utilities") pod "a8a980d8-c735-4920-a9e3-ee6eb98b00b1" (UID: "a8a980d8-c735-4920-a9e3-ee6eb98b00b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.071421 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-kube-api-access-g8fl5" (OuterVolumeSpecName: "kube-api-access-g8fl5") pod "a8a980d8-c735-4920-a9e3-ee6eb98b00b1" (UID: "a8a980d8-c735-4920-a9e3-ee6eb98b00b1"). InnerVolumeSpecName "kube-api-access-g8fl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.113438 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8a980d8-c735-4920-a9e3-ee6eb98b00b1" (UID: "a8a980d8-c735-4920-a9e3-ee6eb98b00b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.164555 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8fl5\" (UniqueName: \"kubernetes.io/projected/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-kube-api-access-g8fl5\") on node \"crc\" DevicePath \"\"" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.164591 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.164603 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8a980d8-c735-4920-a9e3-ee6eb98b00b1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.907683 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pf4pd" Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.951996 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pf4pd"] Feb 20 00:18:51 crc kubenswrapper[4750]: I0220 00:18:51.962205 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pf4pd"] Feb 20 00:18:52 crc kubenswrapper[4750]: I0220 00:18:52.583285 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" path="/var/lib/kubelet/pods/a8a980d8-c735-4920-a9e3-ee6eb98b00b1/volumes" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.347002 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c"] Feb 20 00:18:54 crc kubenswrapper[4750]: E0220 00:18:54.347206 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="extract-utilities" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.347217 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="extract-utilities" Feb 20 00:18:54 crc kubenswrapper[4750]: E0220 00:18:54.347230 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="registry-server" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.347236 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="registry-server" Feb 20 00:18:54 crc kubenswrapper[4750]: E0220 00:18:54.347246 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="extract-content" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.347251 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="extract-content" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.347336 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8a980d8-c735-4920-a9e3-ee6eb98b00b1" containerName="registry-server" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.347962 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.350139 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.407131 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c"] Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.410642 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.410717 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxqqq\" (UniqueName: \"kubernetes.io/projected/744a8b27-0662-4e4f-8d91-1a2740182b2f-kube-api-access-hxqqq\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.410751 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.512557 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.513231 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxqqq\" (UniqueName: \"kubernetes.io/projected/744a8b27-0662-4e4f-8d91-1a2740182b2f-kube-api-access-hxqqq\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.513420 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.513456 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.513854 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.538851 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxqqq\" (UniqueName: \"kubernetes.io/projected/744a8b27-0662-4e4f-8d91-1a2740182b2f-kube-api-access-hxqqq\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.676420 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:18:54 crc kubenswrapper[4750]: I0220 00:18:54.981622 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c"] Feb 20 00:18:54 crc kubenswrapper[4750]: W0220 00:18:54.992138 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod744a8b27_0662_4e4f_8d91_1a2740182b2f.slice/crio-39686ff6a38df9a1fd7a050fb479b91fd46bae58515890a013c6290730100e17 WatchSource:0}: Error finding container 39686ff6a38df9a1fd7a050fb479b91fd46bae58515890a013c6290730100e17: Status 404 returned error can't find the container with id 39686ff6a38df9a1fd7a050fb479b91fd46bae58515890a013c6290730100e17 Feb 20 00:18:55 crc kubenswrapper[4750]: I0220 00:18:55.935400 4750 generic.go:334] "Generic (PLEG): container finished" podID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerID="d4a475b960fda2db9cd27e52642b6504766446be3adf24a79a9b8001d622aef7" exitCode=0 Feb 20 00:18:55 crc kubenswrapper[4750]: I0220 00:18:55.935477 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" event={"ID":"744a8b27-0662-4e4f-8d91-1a2740182b2f","Type":"ContainerDied","Data":"d4a475b960fda2db9cd27e52642b6504766446be3adf24a79a9b8001d622aef7"} Feb 20 00:18:55 crc kubenswrapper[4750]: I0220 00:18:55.935678 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" event={"ID":"744a8b27-0662-4e4f-8d91-1a2740182b2f","Type":"ContainerStarted","Data":"39686ff6a38df9a1fd7a050fb479b91fd46bae58515890a013c6290730100e17"} Feb 20 00:18:55 crc kubenswrapper[4750]: I0220 00:18:55.937856 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 00:18:57 crc kubenswrapper[4750]: I0220 00:18:57.954002 4750 generic.go:334] "Generic (PLEG): container finished" podID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerID="55754ea9dd357e0efdc2c971d10b3c5bb3c00085b53a941758163ce354b5c06c" exitCode=0 Feb 20 00:18:57 crc kubenswrapper[4750]: I0220 00:18:57.954174 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" event={"ID":"744a8b27-0662-4e4f-8d91-1a2740182b2f","Type":"ContainerDied","Data":"55754ea9dd357e0efdc2c971d10b3c5bb3c00085b53a941758163ce354b5c06c"} Feb 20 00:18:58 crc kubenswrapper[4750]: I0220 00:18:58.967220 4750 generic.go:334] "Generic (PLEG): container finished" podID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerID="3ae9d02f646f843e1defb4c333da1867f7987877f69090b41dea09cd1bc56f31" exitCode=0 Feb 20 00:18:58 crc kubenswrapper[4750]: I0220 00:18:58.967480 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" event={"ID":"744a8b27-0662-4e4f-8d91-1a2740182b2f","Type":"ContainerDied","Data":"3ae9d02f646f843e1defb4c333da1867f7987877f69090b41dea09cd1bc56f31"} Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.313499 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.333164 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt"] Feb 20 00:19:00 crc kubenswrapper[4750]: E0220 00:19:00.333431 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerName="pull" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.333445 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerName="pull" Feb 20 00:19:00 crc kubenswrapper[4750]: E0220 00:19:00.333463 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerName="extract" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.333475 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerName="extract" Feb 20 00:19:00 crc kubenswrapper[4750]: E0220 00:19:00.333492 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerName="util" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.333499 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerName="util" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.333610 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="744a8b27-0662-4e4f-8d91-1a2740182b2f" containerName="extract" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.336491 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.350924 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt"] Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.433432 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxqqq\" (UniqueName: \"kubernetes.io/projected/744a8b27-0662-4e4f-8d91-1a2740182b2f-kube-api-access-hxqqq\") pod \"744a8b27-0662-4e4f-8d91-1a2740182b2f\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.433531 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-bundle\") pod \"744a8b27-0662-4e4f-8d91-1a2740182b2f\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.433562 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-util\") pod \"744a8b27-0662-4e4f-8d91-1a2740182b2f\" (UID: \"744a8b27-0662-4e4f-8d91-1a2740182b2f\") " Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.433787 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.433872 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jml4r\" (UniqueName: \"kubernetes.io/projected/152dd755-10c1-482a-872d-31868f80c26f-kube-api-access-jml4r\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.433915 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.436869 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-bundle" (OuterVolumeSpecName: "bundle") pod "744a8b27-0662-4e4f-8d91-1a2740182b2f" (UID: "744a8b27-0662-4e4f-8d91-1a2740182b2f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.443183 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/744a8b27-0662-4e4f-8d91-1a2740182b2f-kube-api-access-hxqqq" (OuterVolumeSpecName: "kube-api-access-hxqqq") pod "744a8b27-0662-4e4f-8d91-1a2740182b2f" (UID: "744a8b27-0662-4e4f-8d91-1a2740182b2f"). InnerVolumeSpecName "kube-api-access-hxqqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.461827 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-util" (OuterVolumeSpecName: "util") pod "744a8b27-0662-4e4f-8d91-1a2740182b2f" (UID: "744a8b27-0662-4e4f-8d91-1a2740182b2f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.534812 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jml4r\" (UniqueName: \"kubernetes.io/projected/152dd755-10c1-482a-872d-31868f80c26f-kube-api-access-jml4r\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.534969 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.535086 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.535267 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxqqq\" (UniqueName: \"kubernetes.io/projected/744a8b27-0662-4e4f-8d91-1a2740182b2f-kube-api-access-hxqqq\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.535307 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.535333 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/744a8b27-0662-4e4f-8d91-1a2740182b2f-util\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.535773 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.535854 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.566300 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jml4r\" (UniqueName: \"kubernetes.io/projected/152dd755-10c1-482a-872d-31868f80c26f-kube-api-access-jml4r\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.662743 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.926052 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt"] Feb 20 00:19:00 crc kubenswrapper[4750]: W0220 00:19:00.927030 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod152dd755_10c1_482a_872d_31868f80c26f.slice/crio-f96134ccd5d06daa817fa4de2ffdd7c3403921c0762e32256d03beb749cd3ea8 WatchSource:0}: Error finding container f96134ccd5d06daa817fa4de2ffdd7c3403921c0762e32256d03beb749cd3ea8: Status 404 returned error can't find the container with id f96134ccd5d06daa817fa4de2ffdd7c3403921c0762e32256d03beb749cd3ea8 Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.983813 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" event={"ID":"152dd755-10c1-482a-872d-31868f80c26f","Type":"ContainerStarted","Data":"f96134ccd5d06daa817fa4de2ffdd7c3403921c0762e32256d03beb749cd3ea8"} Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.987098 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" event={"ID":"744a8b27-0662-4e4f-8d91-1a2740182b2f","Type":"ContainerDied","Data":"39686ff6a38df9a1fd7a050fb479b91fd46bae58515890a013c6290730100e17"} Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.987539 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39686ff6a38df9a1fd7a050fb479b91fd46bae58515890a013c6290730100e17" Feb 20 00:19:00 crc kubenswrapper[4750]: I0220 00:19:00.987429 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.124342 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv"] Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.125537 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.133615 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv"] Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.245762 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.245831 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.245889 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n46bj\" (UniqueName: \"kubernetes.io/projected/ebc81384-b7b6-4011-9040-d5655369377f-kube-api-access-n46bj\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.346946 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n46bj\" (UniqueName: \"kubernetes.io/projected/ebc81384-b7b6-4011-9040-d5655369377f-kube-api-access-n46bj\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.347032 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.347077 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.347917 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.348346 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.366757 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n46bj\" (UniqueName: \"kubernetes.io/projected/ebc81384-b7b6-4011-9040-d5655369377f-kube-api-access-n46bj\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.548837 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:01 crc kubenswrapper[4750]: I0220 00:19:01.766050 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv"] Feb 20 00:19:02 crc kubenswrapper[4750]: I0220 00:19:02.004566 4750 generic.go:334] "Generic (PLEG): container finished" podID="ebc81384-b7b6-4011-9040-d5655369377f" containerID="0cb12cc938c9944a0c8fb65e1cd13613d204dbdb48f84117489cacfab9611dc2" exitCode=0 Feb 20 00:19:02 crc kubenswrapper[4750]: I0220 00:19:02.004631 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" event={"ID":"ebc81384-b7b6-4011-9040-d5655369377f","Type":"ContainerDied","Data":"0cb12cc938c9944a0c8fb65e1cd13613d204dbdb48f84117489cacfab9611dc2"} Feb 20 00:19:02 crc kubenswrapper[4750]: I0220 00:19:02.005099 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" event={"ID":"ebc81384-b7b6-4011-9040-d5655369377f","Type":"ContainerStarted","Data":"3322a9c5e8fa67daf9fa762b1664a65ab600b748582dad7bd2aa3f6cb75c556d"} Feb 20 00:19:02 crc kubenswrapper[4750]: I0220 00:19:02.007648 4750 generic.go:334] "Generic (PLEG): container finished" podID="152dd755-10c1-482a-872d-31868f80c26f" containerID="76ba7cd2d87ea6825365e45f1f9d697112851b5aef682911d05afa34aceada80" exitCode=0 Feb 20 00:19:02 crc kubenswrapper[4750]: I0220 00:19:02.007718 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" event={"ID":"152dd755-10c1-482a-872d-31868f80c26f","Type":"ContainerDied","Data":"76ba7cd2d87ea6825365e45f1f9d697112851b5aef682911d05afa34aceada80"} Feb 20 00:19:04 crc kubenswrapper[4750]: I0220 00:19:04.020550 4750 generic.go:334] "Generic (PLEG): container finished" podID="ebc81384-b7b6-4011-9040-d5655369377f" containerID="96edc52237fb409feaae3388b5064c91f5d09afbc4d7fb99f03314aaf47d80dd" exitCode=0 Feb 20 00:19:04 crc kubenswrapper[4750]: I0220 00:19:04.021562 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" event={"ID":"ebc81384-b7b6-4011-9040-d5655369377f","Type":"ContainerDied","Data":"96edc52237fb409feaae3388b5064c91f5d09afbc4d7fb99f03314aaf47d80dd"} Feb 20 00:19:04 crc kubenswrapper[4750]: I0220 00:19:04.024656 4750 generic.go:334] "Generic (PLEG): container finished" podID="152dd755-10c1-482a-872d-31868f80c26f" containerID="a3ea17168ca0c01f9b7db6320d66fed6a48b8a2c97af5bdfc12deba391287dcc" exitCode=0 Feb 20 00:19:04 crc kubenswrapper[4750]: I0220 00:19:04.024698 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" event={"ID":"152dd755-10c1-482a-872d-31868f80c26f","Type":"ContainerDied","Data":"a3ea17168ca0c01f9b7db6320d66fed6a48b8a2c97af5bdfc12deba391287dcc"} Feb 20 00:19:05 crc kubenswrapper[4750]: I0220 00:19:05.031206 4750 generic.go:334] "Generic (PLEG): container finished" podID="ebc81384-b7b6-4011-9040-d5655369377f" containerID="e8616350498192487d578cd460526749853feedfce45165e8721f5c273966ae1" exitCode=0 Feb 20 00:19:05 crc kubenswrapper[4750]: I0220 00:19:05.031516 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" event={"ID":"ebc81384-b7b6-4011-9040-d5655369377f","Type":"ContainerDied","Data":"e8616350498192487d578cd460526749853feedfce45165e8721f5c273966ae1"} Feb 20 00:19:05 crc kubenswrapper[4750]: I0220 00:19:05.033678 4750 generic.go:334] "Generic (PLEG): container finished" podID="152dd755-10c1-482a-872d-31868f80c26f" containerID="a177415e9145e3b50899c174bebec9d73e90ecc52e676bb46354d92a227e6e8c" exitCode=0 Feb 20 00:19:05 crc kubenswrapper[4750]: I0220 00:19:05.033700 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" event={"ID":"152dd755-10c1-482a-872d-31868f80c26f","Type":"ContainerDied","Data":"a177415e9145e3b50899c174bebec9d73e90ecc52e676bb46354d92a227e6e8c"} Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.454972 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.508257 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-util\") pod \"152dd755-10c1-482a-872d-31868f80c26f\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.508298 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jml4r\" (UniqueName: \"kubernetes.io/projected/152dd755-10c1-482a-872d-31868f80c26f-kube-api-access-jml4r\") pod \"152dd755-10c1-482a-872d-31868f80c26f\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.508342 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-bundle\") pod \"152dd755-10c1-482a-872d-31868f80c26f\" (UID: \"152dd755-10c1-482a-872d-31868f80c26f\") " Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.508987 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-bundle" (OuterVolumeSpecName: "bundle") pod "152dd755-10c1-482a-872d-31868f80c26f" (UID: "152dd755-10c1-482a-872d-31868f80c26f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.517268 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/152dd755-10c1-482a-872d-31868f80c26f-kube-api-access-jml4r" (OuterVolumeSpecName: "kube-api-access-jml4r") pod "152dd755-10c1-482a-872d-31868f80c26f" (UID: "152dd755-10c1-482a-872d-31868f80c26f"). InnerVolumeSpecName "kube-api-access-jml4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.601773 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-util" (OuterVolumeSpecName: "util") pod "152dd755-10c1-482a-872d-31868f80c26f" (UID: "152dd755-10c1-482a-872d-31868f80c26f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.602618 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.609719 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n46bj\" (UniqueName: \"kubernetes.io/projected/ebc81384-b7b6-4011-9040-d5655369377f-kube-api-access-n46bj\") pod \"ebc81384-b7b6-4011-9040-d5655369377f\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.609812 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-bundle\") pod \"ebc81384-b7b6-4011-9040-d5655369377f\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.609838 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-util\") pod \"ebc81384-b7b6-4011-9040-d5655369377f\" (UID: \"ebc81384-b7b6-4011-9040-d5655369377f\") " Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.610081 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-util\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.610099 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jml4r\" (UniqueName: \"kubernetes.io/projected/152dd755-10c1-482a-872d-31868f80c26f-kube-api-access-jml4r\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.610110 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/152dd755-10c1-482a-872d-31868f80c26f-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.610667 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-bundle" (OuterVolumeSpecName: "bundle") pod "ebc81384-b7b6-4011-9040-d5655369377f" (UID: "ebc81384-b7b6-4011-9040-d5655369377f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.613900 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebc81384-b7b6-4011-9040-d5655369377f-kube-api-access-n46bj" (OuterVolumeSpecName: "kube-api-access-n46bj") pod "ebc81384-b7b6-4011-9040-d5655369377f" (UID: "ebc81384-b7b6-4011-9040-d5655369377f"). InnerVolumeSpecName "kube-api-access-n46bj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.628408 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-util" (OuterVolumeSpecName: "util") pod "ebc81384-b7b6-4011-9040-d5655369377f" (UID: "ebc81384-b7b6-4011-9040-d5655369377f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.711362 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n46bj\" (UniqueName: \"kubernetes.io/projected/ebc81384-b7b6-4011-9040-d5655369377f-kube-api-access-n46bj\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.711601 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:06 crc kubenswrapper[4750]: I0220 00:19:06.711708 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebc81384-b7b6-4011-9040-d5655369377f-util\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:07 crc kubenswrapper[4750]: I0220 00:19:07.044353 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" event={"ID":"ebc81384-b7b6-4011-9040-d5655369377f","Type":"ContainerDied","Data":"3322a9c5e8fa67daf9fa762b1664a65ab600b748582dad7bd2aa3f6cb75c556d"} Feb 20 00:19:07 crc kubenswrapper[4750]: I0220 00:19:07.044390 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3322a9c5e8fa67daf9fa762b1664a65ab600b748582dad7bd2aa3f6cb75c556d" Feb 20 00:19:07 crc kubenswrapper[4750]: I0220 00:19:07.044402 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv" Feb 20 00:19:07 crc kubenswrapper[4750]: I0220 00:19:07.055660 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" event={"ID":"152dd755-10c1-482a-872d-31868f80c26f","Type":"ContainerDied","Data":"f96134ccd5d06daa817fa4de2ffdd7c3403921c0762e32256d03beb749cd3ea8"} Feb 20 00:19:07 crc kubenswrapper[4750]: I0220 00:19:07.055693 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f96134ccd5d06daa817fa4de2ffdd7c3403921c0762e32256d03beb749cd3ea8" Feb 20 00:19:07 crc kubenswrapper[4750]: I0220 00:19:07.055710 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.275403 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94"] Feb 20 00:19:09 crc kubenswrapper[4750]: E0220 00:19:09.275940 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc81384-b7b6-4011-9040-d5655369377f" containerName="pull" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.275956 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc81384-b7b6-4011-9040-d5655369377f" containerName="pull" Feb 20 00:19:09 crc kubenswrapper[4750]: E0220 00:19:09.275974 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152dd755-10c1-482a-872d-31868f80c26f" containerName="util" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.275981 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="152dd755-10c1-482a-872d-31868f80c26f" containerName="util" Feb 20 00:19:09 crc kubenswrapper[4750]: E0220 00:19:09.275994 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152dd755-10c1-482a-872d-31868f80c26f" containerName="pull" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.276003 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="152dd755-10c1-482a-872d-31868f80c26f" containerName="pull" Feb 20 00:19:09 crc kubenswrapper[4750]: E0220 00:19:09.276015 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc81384-b7b6-4011-9040-d5655369377f" containerName="extract" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.276022 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc81384-b7b6-4011-9040-d5655369377f" containerName="extract" Feb 20 00:19:09 crc kubenswrapper[4750]: E0220 00:19:09.276031 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="152dd755-10c1-482a-872d-31868f80c26f" containerName="extract" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.276038 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="152dd755-10c1-482a-872d-31868f80c26f" containerName="extract" Feb 20 00:19:09 crc kubenswrapper[4750]: E0220 00:19:09.276049 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc81384-b7b6-4011-9040-d5655369377f" containerName="util" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.276055 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc81384-b7b6-4011-9040-d5655369377f" containerName="util" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.276183 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="152dd755-10c1-482a-872d-31868f80c26f" containerName="extract" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.276200 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebc81384-b7b6-4011-9040-d5655369377f" containerName="extract" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.276643 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.286438 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.289328 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.290306 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-zfvvc" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.290532 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.368024 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.369274 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.372626 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.380333 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.409139 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.410073 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.414536 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.414617 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-vbwhl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.425264 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.426141 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.434325 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.440819 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tqvl\" (UniqueName: \"kubernetes.io/projected/359c9e5a-f1ba-4513-9d7a-0903ae871923-kube-api-access-4tqvl\") pod \"obo-prometheus-operator-68bc856cb9-6kn94\" (UID: \"359c9e5a-f1ba-4513-9d7a-0903ae871923\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.450360 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.500792 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-gsxjb"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.504008 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.505944 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-dvkjg" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.506442 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.517914 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-gsxjb"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542008 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tqvl\" (UniqueName: \"kubernetes.io/projected/359c9e5a-f1ba-4513-9d7a-0903ae871923-kube-api-access-4tqvl\") pod \"obo-prometheus-operator-68bc856cb9-6kn94\" (UID: \"359c9e5a-f1ba-4513-9d7a-0903ae871923\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542090 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df8960fb-1d2c-4857-988d-f51ea636c079-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89\" (UID: \"df8960fb-1d2c-4857-988d-f51ea636c079\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542149 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542174 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542826 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnw7w\" (UniqueName: \"kubernetes.io/projected/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-kube-api-access-tnw7w\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542880 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3d27f900-e106-4a95-b60c-b75c81bf21e6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r\" (UID: \"3d27f900-e106-4a95-b60c-b75c81bf21e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542959 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3d27f900-e106-4a95-b60c-b75c81bf21e6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r\" (UID: \"3d27f900-e106-4a95-b60c-b75c81bf21e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.542992 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df8960fb-1d2c-4857-988d-f51ea636c079-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89\" (UID: \"df8960fb-1d2c-4857-988d-f51ea636c079\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.565943 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tqvl\" (UniqueName: \"kubernetes.io/projected/359c9e5a-f1ba-4513-9d7a-0903ae871923-kube-api-access-4tqvl\") pod \"obo-prometheus-operator-68bc856cb9-6kn94\" (UID: \"359c9e5a-f1ba-4513-9d7a-0903ae871923\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.595164 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.600820 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-5vmsh"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.601977 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.604517 4750 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-6kxzd" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.608816 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-5vmsh"] Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643701 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnw7w\" (UniqueName: \"kubernetes.io/projected/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-kube-api-access-tnw7w\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643767 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3d27f900-e106-4a95-b60c-b75c81bf21e6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r\" (UID: \"3d27f900-e106-4a95-b60c-b75c81bf21e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643808 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3d27f900-e106-4a95-b60c-b75c81bf21e6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r\" (UID: \"3d27f900-e106-4a95-b60c-b75c81bf21e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643827 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df8960fb-1d2c-4857-988d-f51ea636c079-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89\" (UID: \"df8960fb-1d2c-4857-988d-f51ea636c079\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643857 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df8960fb-1d2c-4857-988d-f51ea636c079-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89\" (UID: \"df8960fb-1d2c-4857-988d-f51ea636c079\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643881 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njbrd\" (UniqueName: \"kubernetes.io/projected/14f2560a-b7fb-491c-86ed-7b1e6fd77f5e-kube-api-access-njbrd\") pod \"observability-operator-59bdc8b94-gsxjb\" (UID: \"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e\") " pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643904 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643921 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.643941 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/14f2560a-b7fb-491c-86ed-7b1e6fd77f5e-observability-operator-tls\") pod \"observability-operator-59bdc8b94-gsxjb\" (UID: \"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e\") " pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.646790 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.647507 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.649784 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3d27f900-e106-4a95-b60c-b75c81bf21e6-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r\" (UID: \"3d27f900-e106-4a95-b60c-b75c81bf21e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.649971 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3d27f900-e106-4a95-b60c-b75c81bf21e6-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r\" (UID: \"3d27f900-e106-4a95-b60c-b75c81bf21e6\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.655483 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df8960fb-1d2c-4857-988d-f51ea636c079-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89\" (UID: \"df8960fb-1d2c-4857-988d-f51ea636c079\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.658721 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df8960fb-1d2c-4857-988d-f51ea636c079-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89\" (UID: \"df8960fb-1d2c-4857-988d-f51ea636c079\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.681773 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnw7w\" (UniqueName: \"kubernetes.io/projected/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-kube-api-access-tnw7w\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.683938 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.725493 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.741635 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.747913 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc4519f1-96b2-4f48-a78c-10c8fda991a1-openshift-service-ca\") pod \"perses-operator-5bf474d74f-5vmsh\" (UID: \"fc4519f1-96b2-4f48-a78c-10c8fda991a1\") " pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.747987 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpw47\" (UniqueName: \"kubernetes.io/projected/fc4519f1-96b2-4f48-a78c-10c8fda991a1-kube-api-access-qpw47\") pod \"perses-operator-5bf474d74f-5vmsh\" (UID: \"fc4519f1-96b2-4f48-a78c-10c8fda991a1\") " pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.748043 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njbrd\" (UniqueName: \"kubernetes.io/projected/14f2560a-b7fb-491c-86ed-7b1e6fd77f5e-kube-api-access-njbrd\") pod \"observability-operator-59bdc8b94-gsxjb\" (UID: \"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e\") " pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.748070 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/14f2560a-b7fb-491c-86ed-7b1e6fd77f5e-observability-operator-tls\") pod \"observability-operator-59bdc8b94-gsxjb\" (UID: \"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e\") " pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.751089 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/14f2560a-b7fb-491c-86ed-7b1e6fd77f5e-observability-operator-tls\") pod \"observability-operator-59bdc8b94-gsxjb\" (UID: \"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e\") " pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.772996 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njbrd\" (UniqueName: \"kubernetes.io/projected/14f2560a-b7fb-491c-86ed-7b1e6fd77f5e-kube-api-access-njbrd\") pod \"observability-operator-59bdc8b94-gsxjb\" (UID: \"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e\") " pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.817216 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.848619 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc4519f1-96b2-4f48-a78c-10c8fda991a1-openshift-service-ca\") pod \"perses-operator-5bf474d74f-5vmsh\" (UID: \"fc4519f1-96b2-4f48-a78c-10c8fda991a1\") " pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.848685 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpw47\" (UniqueName: \"kubernetes.io/projected/fc4519f1-96b2-4f48-a78c-10c8fda991a1-kube-api-access-qpw47\") pod \"perses-operator-5bf474d74f-5vmsh\" (UID: \"fc4519f1-96b2-4f48-a78c-10c8fda991a1\") " pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.849714 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/fc4519f1-96b2-4f48-a78c-10c8fda991a1-openshift-service-ca\") pod \"perses-operator-5bf474d74f-5vmsh\" (UID: \"fc4519f1-96b2-4f48-a78c-10c8fda991a1\") " pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:09 crc kubenswrapper[4750]: I0220 00:19:09.876293 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpw47\" (UniqueName: \"kubernetes.io/projected/fc4519f1-96b2-4f48-a78c-10c8fda991a1-kube-api-access-qpw47\") pod \"perses-operator-5bf474d74f-5vmsh\" (UID: \"fc4519f1-96b2-4f48-a78c-10c8fda991a1\") " pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.011769 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.013565 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94"] Feb 20 00:19:10 crc kubenswrapper[4750]: W0220 00:19:10.047262 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359c9e5a_f1ba_4513_9d7a_0903ae871923.slice/crio-0d5b8f2f5fe00f3c06c70233c3348aae229f4c88ca1fd2dfcc1186b1ed8729ea WatchSource:0}: Error finding container 0d5b8f2f5fe00f3c06c70233c3348aae229f4c88ca1fd2dfcc1186b1ed8729ea: Status 404 returned error can't find the container with id 0d5b8f2f5fe00f3c06c70233c3348aae229f4c88ca1fd2dfcc1186b1ed8729ea Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.097609 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89"] Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.098746 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" event={"ID":"359c9e5a-f1ba-4513-9d7a-0903ae871923","Type":"ContainerStarted","Data":"0d5b8f2f5fe00f3c06c70233c3348aae229f4c88ca1fd2dfcc1186b1ed8729ea"} Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.174853 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r"] Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.292900 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl"] Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.332476 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-5vmsh"] Feb 20 00:19:10 crc kubenswrapper[4750]: W0220 00:19:10.339332 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc4519f1_96b2_4f48_a78c_10c8fda991a1.slice/crio-bea7496bcb7c4d93fb6c7668d6a4d7cca2db36bfb47df0e5610a1ef9762ce1aa WatchSource:0}: Error finding container bea7496bcb7c4d93fb6c7668d6a4d7cca2db36bfb47df0e5610a1ef9762ce1aa: Status 404 returned error can't find the container with id bea7496bcb7c4d93fb6c7668d6a4d7cca2db36bfb47df0e5610a1ef9762ce1aa Feb 20 00:19:10 crc kubenswrapper[4750]: I0220 00:19:10.448938 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-gsxjb"] Feb 20 00:19:11 crc kubenswrapper[4750]: I0220 00:19:11.109540 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" event={"ID":"3d27f900-e106-4a95-b60c-b75c81bf21e6","Type":"ContainerStarted","Data":"84a67883c42b71174c156330edef996cfcfe25719e56002ef5d5e329134f552e"} Feb 20 00:19:11 crc kubenswrapper[4750]: I0220 00:19:11.111073 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" event={"ID":"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e","Type":"ContainerStarted","Data":"740fa0e50bc8edd0113f5ac532f549093de52f22ce44efd6bb6520f97c0a3d62"} Feb 20 00:19:11 crc kubenswrapper[4750]: I0220 00:19:11.115571 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" event={"ID":"df8960fb-1d2c-4857-988d-f51ea636c079","Type":"ContainerStarted","Data":"464b32df78b0fee42097ffe05dfc3c7d1a1b47bbf493dfe24773dcc55366d33d"} Feb 20 00:19:11 crc kubenswrapper[4750]: I0220 00:19:11.118058 4750 generic.go:334] "Generic (PLEG): container finished" podID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerID="88c5b95a7b5eee8dbc0af50c56d4b3dac8fc867a473d66aedaa5037275a1ce2a" exitCode=0 Feb 20 00:19:11 crc kubenswrapper[4750]: I0220 00:19:11.118143 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" event={"ID":"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf","Type":"ContainerDied","Data":"88c5b95a7b5eee8dbc0af50c56d4b3dac8fc867a473d66aedaa5037275a1ce2a"} Feb 20 00:19:11 crc kubenswrapper[4750]: I0220 00:19:11.118173 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" event={"ID":"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf","Type":"ContainerStarted","Data":"4c3c4bf19f035ee5a2d60463edcc1c8cc54c5dd86c50e0ad170bd5814bcb7e49"} Feb 20 00:19:11 crc kubenswrapper[4750]: I0220 00:19:11.119772 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" event={"ID":"fc4519f1-96b2-4f48-a78c-10c8fda991a1","Type":"ContainerStarted","Data":"bea7496bcb7c4d93fb6c7668d6a4d7cca2db36bfb47df0e5610a1ef9762ce1aa"} Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.012953 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-777ffcd498-dw9xw"] Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.013915 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.017048 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.017352 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-kgvw8" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.017600 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.018040 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.035027 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-777ffcd498-dw9xw"] Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.155188 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m2vv\" (UniqueName: \"kubernetes.io/projected/5a268689-ad0d-493c-958e-d201942ce1ef-kube-api-access-9m2vv\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.155254 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5a268689-ad0d-493c-958e-d201942ce1ef-webhook-cert\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.155290 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5a268689-ad0d-493c-958e-d201942ce1ef-apiservice-cert\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.256705 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5a268689-ad0d-493c-958e-d201942ce1ef-webhook-cert\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.256769 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5a268689-ad0d-493c-958e-d201942ce1ef-apiservice-cert\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.256817 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m2vv\" (UniqueName: \"kubernetes.io/projected/5a268689-ad0d-493c-958e-d201942ce1ef-kube-api-access-9m2vv\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.262608 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5a268689-ad0d-493c-958e-d201942ce1ef-webhook-cert\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.263215 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5a268689-ad0d-493c-958e-d201942ce1ef-apiservice-cert\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.282623 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m2vv\" (UniqueName: \"kubernetes.io/projected/5a268689-ad0d-493c-958e-d201942ce1ef-kube-api-access-9m2vv\") pod \"elastic-operator-777ffcd498-dw9xw\" (UID: \"5a268689-ad0d-493c-958e-d201942ce1ef\") " pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:18 crc kubenswrapper[4750]: I0220 00:19:18.332357 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.012558 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-dlmqf"] Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.015057 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.016839 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-n8jkd" Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.017256 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-dlmqf"] Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.107435 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8lb6\" (UniqueName: \"kubernetes.io/projected/c4b5536e-4184-4e57-be5a-dc42036db7de-kube-api-access-r8lb6\") pod \"interconnect-operator-5bb49f789d-dlmqf\" (UID: \"c4b5536e-4184-4e57-be5a-dc42036db7de\") " pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.208316 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8lb6\" (UniqueName: \"kubernetes.io/projected/c4b5536e-4184-4e57-be5a-dc42036db7de-kube-api-access-r8lb6\") pod \"interconnect-operator-5bb49f789d-dlmqf\" (UID: \"c4b5536e-4184-4e57-be5a-dc42036db7de\") " pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.235873 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8lb6\" (UniqueName: \"kubernetes.io/projected/c4b5536e-4184-4e57-be5a-dc42036db7de-kube-api-access-r8lb6\") pod \"interconnect-operator-5bb49f789d-dlmqf\" (UID: \"c4b5536e-4184-4e57-be5a-dc42036db7de\") " pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.341382 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" Feb 20 00:19:22 crc kubenswrapper[4750]: I0220 00:19:22.881052 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-777ffcd498-dw9xw"] Feb 20 00:19:22 crc kubenswrapper[4750]: W0220 00:19:22.891056 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a268689_ad0d_493c_958e_d201942ce1ef.slice/crio-ec487b36e8ff3d3c622de46ef829e0f64f78ea0481a71f2f5180d3ea16008c17 WatchSource:0}: Error finding container ec487b36e8ff3d3c622de46ef829e0f64f78ea0481a71f2f5180d3ea16008c17: Status 404 returned error can't find the container with id ec487b36e8ff3d3c622de46ef829e0f64f78ea0481a71f2f5180d3ea16008c17 Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.092361 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-dlmqf"] Feb 20 00:19:23 crc kubenswrapper[4750]: W0220 00:19:23.099817 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4b5536e_4184_4e57_be5a_dc42036db7de.slice/crio-5bb5b9446bc8244f03a05835235401236c1ef3ad6c2aabfa70b2fc41e5cf26be WatchSource:0}: Error finding container 5bb5b9446bc8244f03a05835235401236c1ef3ad6c2aabfa70b2fc41e5cf26be: Status 404 returned error can't find the container with id 5bb5b9446bc8244f03a05835235401236c1ef3ad6c2aabfa70b2fc41e5cf26be Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.215922 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" event={"ID":"5a268689-ad0d-493c-958e-d201942ce1ef","Type":"ContainerStarted","Data":"ec487b36e8ff3d3c622de46ef829e0f64f78ea0481a71f2f5180d3ea16008c17"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.217719 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" event={"ID":"14f2560a-b7fb-491c-86ed-7b1e6fd77f5e","Type":"ContainerStarted","Data":"4dba5ad26db96894571251e3e2b5f70dd101cac8bf5305dd3ea09f02202046b3"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.218098 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.219200 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" event={"ID":"df8960fb-1d2c-4857-988d-f51ea636c079","Type":"ContainerStarted","Data":"cb4458d5148497c646bcd406774dbdfecdc08c87df795cd7113a3f96f0df9c68"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.220059 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.221240 4750 generic.go:334] "Generic (PLEG): container finished" podID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerID="58962d585d9a39aca3a8fd50bf4ce85fcedad99abc903741685cefd77a5b794d" exitCode=0 Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.221274 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" event={"ID":"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf","Type":"ContainerDied","Data":"58962d585d9a39aca3a8fd50bf4ce85fcedad99abc903741685cefd77a5b794d"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.223317 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" event={"ID":"359c9e5a-f1ba-4513-9d7a-0903ae871923","Type":"ContainerStarted","Data":"6c36fcc61e4c687b0b571238e2a342f81ad064de63791f9e35d998d252ee5dfe"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.224941 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" event={"ID":"fc4519f1-96b2-4f48-a78c-10c8fda991a1","Type":"ContainerStarted","Data":"4d25fa96bb05215de0a8399f596589f0f429e6c861bb42fe804c3d7cf9617f1f"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.225075 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.226214 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" event={"ID":"3d27f900-e106-4a95-b60c-b75c81bf21e6","Type":"ContainerStarted","Data":"557672ce7fc67989ad8a0a09bd36df255888dcc48d0b7144df8bca8c28fbb330"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.227412 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" event={"ID":"c4b5536e-4184-4e57-be5a-dc42036db7de","Type":"ContainerStarted","Data":"5bb5b9446bc8244f03a05835235401236c1ef3ad6c2aabfa70b2fc41e5cf26be"} Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.252171 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-gsxjb" podStartSLOduration=2.082848055 podStartE2EDuration="14.252113357s" podCreationTimestamp="2026-02-20 00:19:09 +0000 UTC" firstStartedPulling="2026-02-20 00:19:10.45832348 +0000 UTC m=+634.653159529" lastFinishedPulling="2026-02-20 00:19:22.627588782 +0000 UTC m=+646.822424831" observedRunningTime="2026-02-20 00:19:23.24691371 +0000 UTC m=+647.441749779" watchObservedRunningTime="2026-02-20 00:19:23.252113357 +0000 UTC m=+647.446949436" Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.315701 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r" podStartSLOduration=1.932569006 podStartE2EDuration="14.315681534s" podCreationTimestamp="2026-02-20 00:19:09 +0000 UTC" firstStartedPulling="2026-02-20 00:19:10.194654098 +0000 UTC m=+634.389490147" lastFinishedPulling="2026-02-20 00:19:22.577766616 +0000 UTC m=+646.772602675" observedRunningTime="2026-02-20 00:19:23.315186181 +0000 UTC m=+647.510022230" watchObservedRunningTime="2026-02-20 00:19:23.315681534 +0000 UTC m=+647.510517573" Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.317697 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89" podStartSLOduration=1.807308769 podStartE2EDuration="14.317687355s" podCreationTimestamp="2026-02-20 00:19:09 +0000 UTC" firstStartedPulling="2026-02-20 00:19:10.121260873 +0000 UTC m=+634.316096922" lastFinishedPulling="2026-02-20 00:19:22.631639459 +0000 UTC m=+646.826475508" observedRunningTime="2026-02-20 00:19:23.28277311 +0000 UTC m=+647.477609159" watchObservedRunningTime="2026-02-20 00:19:23.317687355 +0000 UTC m=+647.512523404" Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.337293 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6kn94" podStartSLOduration=1.8224999579999999 podStartE2EDuration="14.337275199s" podCreationTimestamp="2026-02-20 00:19:09 +0000 UTC" firstStartedPulling="2026-02-20 00:19:10.051268878 +0000 UTC m=+634.246104927" lastFinishedPulling="2026-02-20 00:19:22.566044109 +0000 UTC m=+646.760880168" observedRunningTime="2026-02-20 00:19:23.336054807 +0000 UTC m=+647.530890856" watchObservedRunningTime="2026-02-20 00:19:23.337275199 +0000 UTC m=+647.532111248" Feb 20 00:19:23 crc kubenswrapper[4750]: I0220 00:19:23.398722 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" podStartSLOduration=2.108826046 podStartE2EDuration="14.39870202s" podCreationTimestamp="2026-02-20 00:19:09 +0000 UTC" firstStartedPulling="2026-02-20 00:19:10.341764835 +0000 UTC m=+634.536600904" lastFinishedPulling="2026-02-20 00:19:22.631640829 +0000 UTC m=+646.826476878" observedRunningTime="2026-02-20 00:19:23.397684673 +0000 UTC m=+647.592520742" watchObservedRunningTime="2026-02-20 00:19:23.39870202 +0000 UTC m=+647.593538079" Feb 20 00:19:24 crc kubenswrapper[4750]: I0220 00:19:24.237194 4750 generic.go:334] "Generic (PLEG): container finished" podID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerID="63bf786e5feeddee6b284b2bc5a030b8d848bed8ec91bfc917bb50360e4f739f" exitCode=0 Feb 20 00:19:24 crc kubenswrapper[4750]: I0220 00:19:24.237441 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" event={"ID":"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf","Type":"ContainerDied","Data":"63bf786e5feeddee6b284b2bc5a030b8d848bed8ec91bfc917bb50360e4f739f"} Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.611182 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.753881 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-util\") pod \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.753955 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnw7w\" (UniqueName: \"kubernetes.io/projected/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-kube-api-access-tnw7w\") pod \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.753991 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-bundle\") pod \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\" (UID: \"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf\") " Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.755272 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-bundle" (OuterVolumeSpecName: "bundle") pod "3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" (UID: "3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.761309 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-kube-api-access-tnw7w" (OuterVolumeSpecName: "kube-api-access-tnw7w") pod "3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" (UID: "3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf"). InnerVolumeSpecName "kube-api-access-tnw7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.771224 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-util" (OuterVolumeSpecName: "util") pod "3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" (UID: "3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.856061 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnw7w\" (UniqueName: \"kubernetes.io/projected/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-kube-api-access-tnw7w\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.856091 4750 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:25 crc kubenswrapper[4750]: I0220 00:19:25.856100 4750 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf-util\") on node \"crc\" DevicePath \"\"" Feb 20 00:19:26 crc kubenswrapper[4750]: I0220 00:19:26.251667 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" event={"ID":"5a268689-ad0d-493c-958e-d201942ce1ef","Type":"ContainerStarted","Data":"310604a8fb81fcf857eae55ed6f8f51015ff085a20bbdcde98e6612946deefe2"} Feb 20 00:19:26 crc kubenswrapper[4750]: I0220 00:19:26.254424 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" event={"ID":"3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf","Type":"ContainerDied","Data":"4c3c4bf19f035ee5a2d60463edcc1c8cc54c5dd86c50e0ad170bd5814bcb7e49"} Feb 20 00:19:26 crc kubenswrapper[4750]: I0220 00:19:26.254467 4750 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c3c4bf19f035ee5a2d60463edcc1c8cc54c5dd86c50e0ad170bd5814bcb7e49" Feb 20 00:19:26 crc kubenswrapper[4750]: I0220 00:19:26.254487 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl" Feb 20 00:19:26 crc kubenswrapper[4750]: I0220 00:19:26.267283 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-777ffcd498-dw9xw" podStartSLOduration=6.690334253 podStartE2EDuration="9.267265978s" podCreationTimestamp="2026-02-20 00:19:17 +0000 UTC" firstStartedPulling="2026-02-20 00:19:22.894319865 +0000 UTC m=+647.089155904" lastFinishedPulling="2026-02-20 00:19:25.47125157 +0000 UTC m=+649.666087629" observedRunningTime="2026-02-20 00:19:26.265059393 +0000 UTC m=+650.459895442" watchObservedRunningTime="2026-02-20 00:19:26.267265978 +0000 UTC m=+650.462102027" Feb 20 00:19:30 crc kubenswrapper[4750]: I0220 00:19:30.014614 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-5vmsh" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.357426 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 20 00:19:31 crc kubenswrapper[4750]: E0220 00:19:31.357780 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerName="extract" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.357792 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerName="extract" Feb 20 00:19:31 crc kubenswrapper[4750]: E0220 00:19:31.357811 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerName="pull" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.357819 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerName="pull" Feb 20 00:19:31 crc kubenswrapper[4750]: E0220 00:19:31.357834 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerName="util" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.357841 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerName="util" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.358013 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf" containerName="extract" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.359664 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.366098 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.366354 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-p59sp" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.366527 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.366772 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.366962 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.367183 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.367215 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.367317 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.367383 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.380066 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.536842 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.536906 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.536934 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.536953 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.536973 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.536992 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537010 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c82277c8-3f73-4244-a092-0117400628a8-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537089 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537108 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537685 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537710 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537821 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537882 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537938 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.537966 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.640504 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.641066 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.640995 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.641177 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.642713 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.642786 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643180 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.642819 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643247 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643290 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643319 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643364 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643428 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643473 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643500 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643539 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643563 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.643631 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c82277c8-3f73-4244-a092-0117400628a8-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.647210 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/c82277c8-3f73-4244-a092-0117400628a8-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.647533 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.648093 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.648776 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.648937 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.649442 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.650245 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.650761 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.653009 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.653372 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.653533 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.653416 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/c82277c8-3f73-4244-a092-0117400628a8-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"c82277c8-3f73-4244-a092-0117400628a8\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:31 crc kubenswrapper[4750]: I0220 00:19:31.690915 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:19:32 crc kubenswrapper[4750]: I0220 00:19:32.164630 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 20 00:19:32 crc kubenswrapper[4750]: W0220 00:19:32.171390 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc82277c8_3f73_4244_a092_0117400628a8.slice/crio-bd8e3cec9df57ab8054a6be85be60316a50ac9b6e41284c60c1c2b5f4643e993 WatchSource:0}: Error finding container bd8e3cec9df57ab8054a6be85be60316a50ac9b6e41284c60c1c2b5f4643e993: Status 404 returned error can't find the container with id bd8e3cec9df57ab8054a6be85be60316a50ac9b6e41284c60c1c2b5f4643e993 Feb 20 00:19:32 crc kubenswrapper[4750]: I0220 00:19:32.292536 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" event={"ID":"c4b5536e-4184-4e57-be5a-dc42036db7de","Type":"ContainerStarted","Data":"c721c8fda603fde5e04a9c20da02937d4b82e4c2b10e4026536c1b284927f90e"} Feb 20 00:19:32 crc kubenswrapper[4750]: I0220 00:19:32.293514 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c82277c8-3f73-4244-a092-0117400628a8","Type":"ContainerStarted","Data":"bd8e3cec9df57ab8054a6be85be60316a50ac9b6e41284c60c1c2b5f4643e993"} Feb 20 00:19:32 crc kubenswrapper[4750]: I0220 00:19:32.308990 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-dlmqf" podStartSLOduration=2.425967393 podStartE2EDuration="11.308970833s" podCreationTimestamp="2026-02-20 00:19:21 +0000 UTC" firstStartedPulling="2026-02-20 00:19:23.101959249 +0000 UTC m=+647.296795298" lastFinishedPulling="2026-02-20 00:19:31.984962689 +0000 UTC m=+656.179798738" observedRunningTime="2026-02-20 00:19:32.30691153 +0000 UTC m=+656.501747579" watchObservedRunningTime="2026-02-20 00:19:32.308970833 +0000 UTC m=+656.503806882" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.526765 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz"] Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.527869 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.531804 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.532013 4750 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-h9xfc" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.532989 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.542936 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz"] Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.652882 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d1a23300-6416-4388-984a-1b912a76ccbf-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-nxxfz\" (UID: \"d1a23300-6416-4388-984a-1b912a76ccbf\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.653316 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c52r\" (UniqueName: \"kubernetes.io/projected/d1a23300-6416-4388-984a-1b912a76ccbf-kube-api-access-8c52r\") pod \"cert-manager-operator-controller-manager-5586865c96-nxxfz\" (UID: \"d1a23300-6416-4388-984a-1b912a76ccbf\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.754351 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c52r\" (UniqueName: \"kubernetes.io/projected/d1a23300-6416-4388-984a-1b912a76ccbf-kube-api-access-8c52r\") pod \"cert-manager-operator-controller-manager-5586865c96-nxxfz\" (UID: \"d1a23300-6416-4388-984a-1b912a76ccbf\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.754408 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d1a23300-6416-4388-984a-1b912a76ccbf-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-nxxfz\" (UID: \"d1a23300-6416-4388-984a-1b912a76ccbf\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.754906 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d1a23300-6416-4388-984a-1b912a76ccbf-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-nxxfz\" (UID: \"d1a23300-6416-4388-984a-1b912a76ccbf\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.776596 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c52r\" (UniqueName: \"kubernetes.io/projected/d1a23300-6416-4388-984a-1b912a76ccbf-kube-api-access-8c52r\") pod \"cert-manager-operator-controller-manager-5586865c96-nxxfz\" (UID: \"d1a23300-6416-4388-984a-1b912a76ccbf\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:45 crc kubenswrapper[4750]: I0220 00:19:45.851153 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" Feb 20 00:19:55 crc kubenswrapper[4750]: E0220 00:19:55.601312 4750 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.connect.redhat.com/elastic/elasticsearch:7.17.20" Feb 20 00:19:55 crc kubenswrapper[4750]: E0220 00:19:55.602062 4750 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:elastic-internal-init-filesystem,Image:registry.connect.redhat.com/elastic/elasticsearch:7.17.20,Command:[bash -c /mnt/elastic-internal/scripts/prepare-fs.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:HEADLESS_SERVICE_NAME,Value:elasticsearch-es-default,ValueFrom:nil,},EnvVar{Name:PROBE_PASSWORD_PATH,Value:/mnt/elastic-internal/pod-mounted-users/elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:PROBE_USERNAME,Value:elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:READINESS_PROBE_PROTOCOL,Value:https,ValueFrom:nil,},EnvVar{Name:NSS_SDB_USE_CACHE,Value:no,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:downward-api,ReadOnly:true,MountPath:/mnt/elastic-internal/downward-api,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-bin-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-bin-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config,ReadOnly:true,MountPath:/mnt/elastic-internal/elasticsearch-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-config-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-plugins-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-plugins-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-http-certificates,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/http-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-probe-user,ReadOnly:true,MountPath:/mnt/elastic-internal/pod-mounted-users,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-remote-certificate-authorities,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/transport-remote-certs/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-scripts,ReadOnly:true,MountPath:/mnt/elastic-internal/scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-transport-certificates,ReadOnly:true,MountPath:/mnt/elastic-internal/transport-certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-unicast-hosts,ReadOnly:true,MountPath:/mnt/elastic-internal/unicast-hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-xpack-file-realm,ReadOnly:true,MountPath:/mnt/elastic-internal/xpack-file-realm,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-data,ReadOnly:false,MountPath:/usr/share/elasticsearch/data,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-logs,ReadOnly:false,MountPath:/usr/share/elasticsearch/logs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-volume,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod elasticsearch-es-default-0_service-telemetry(c82277c8-3f73-4244-a092-0117400628a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 00:19:55 crc kubenswrapper[4750]: E0220 00:19:55.603230 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="c82277c8-3f73-4244-a092-0117400628a8" Feb 20 00:19:55 crc kubenswrapper[4750]: I0220 00:19:55.916241 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz"] Feb 20 00:19:55 crc kubenswrapper[4750]: W0220 00:19:55.929466 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1a23300_6416_4388_984a_1b912a76ccbf.slice/crio-45a64c57e9690c8b80322be7b6800fffee52aec8ac4552d92fa814b753f8003c WatchSource:0}: Error finding container 45a64c57e9690c8b80322be7b6800fffee52aec8ac4552d92fa814b753f8003c: Status 404 returned error can't find the container with id 45a64c57e9690c8b80322be7b6800fffee52aec8ac4552d92fa814b753f8003c Feb 20 00:19:56 crc kubenswrapper[4750]: I0220 00:19:56.536238 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" event={"ID":"d1a23300-6416-4388-984a-1b912a76ccbf","Type":"ContainerStarted","Data":"45a64c57e9690c8b80322be7b6800fffee52aec8ac4552d92fa814b753f8003c"} Feb 20 00:19:56 crc kubenswrapper[4750]: E0220 00:19:56.538027 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="c82277c8-3f73-4244-a092-0117400628a8" Feb 20 00:19:56 crc kubenswrapper[4750]: I0220 00:19:56.723692 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 20 00:19:56 crc kubenswrapper[4750]: I0220 00:19:56.770562 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 20 00:19:57 crc kubenswrapper[4750]: E0220 00:19:57.546566 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="c82277c8-3f73-4244-a092-0117400628a8" Feb 20 00:19:58 crc kubenswrapper[4750]: E0220 00:19:58.555385 4750 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="c82277c8-3f73-4244-a092-0117400628a8" Feb 20 00:19:59 crc kubenswrapper[4750]: I0220 00:19:59.559956 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" event={"ID":"d1a23300-6416-4388-984a-1b912a76ccbf","Type":"ContainerStarted","Data":"3df3f8e7047a91ea6d588f2eafcb80461dcb433d89d9d3169ffa6be1e42d9c20"} Feb 20 00:19:59 crc kubenswrapper[4750]: I0220 00:19:59.581017 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-nxxfz" podStartSLOduration=11.458725238 podStartE2EDuration="14.580996486s" podCreationTimestamp="2026-02-20 00:19:45 +0000 UTC" firstStartedPulling="2026-02-20 00:19:55.931648277 +0000 UTC m=+680.126484326" lastFinishedPulling="2026-02-20 00:19:59.053919515 +0000 UTC m=+683.248755574" observedRunningTime="2026-02-20 00:19:59.578261136 +0000 UTC m=+683.773097215" watchObservedRunningTime="2026-02-20 00:19:59.580996486 +0000 UTC m=+683.775832535" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.447773 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-lg2cd"] Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.449296 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.450989 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.452023 4750 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-g9s9n" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.454494 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.466875 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-lg2cd"] Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.565916 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk47g\" (UniqueName: \"kubernetes.io/projected/1eeff3f9-8995-450d-af18-d129c823ec09-kube-api-access-dk47g\") pod \"cert-manager-cainjector-5545bd876-lg2cd\" (UID: \"1eeff3f9-8995-450d-af18-d129c823ec09\") " pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.566061 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1eeff3f9-8995-450d-af18-d129c823ec09-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-lg2cd\" (UID: \"1eeff3f9-8995-450d-af18-d129c823ec09\") " pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.668154 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk47g\" (UniqueName: \"kubernetes.io/projected/1eeff3f9-8995-450d-af18-d129c823ec09-kube-api-access-dk47g\") pod \"cert-manager-cainjector-5545bd876-lg2cd\" (UID: \"1eeff3f9-8995-450d-af18-d129c823ec09\") " pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.668226 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1eeff3f9-8995-450d-af18-d129c823ec09-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-lg2cd\" (UID: \"1eeff3f9-8995-450d-af18-d129c823ec09\") " pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.690050 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk47g\" (UniqueName: \"kubernetes.io/projected/1eeff3f9-8995-450d-af18-d129c823ec09-kube-api-access-dk47g\") pod \"cert-manager-cainjector-5545bd876-lg2cd\" (UID: \"1eeff3f9-8995-450d-af18-d129c823ec09\") " pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.695658 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1eeff3f9-8995-450d-af18-d129c823ec09-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-lg2cd\" (UID: \"1eeff3f9-8995-450d-af18-d129c823ec09\") " pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:06 crc kubenswrapper[4750]: I0220 00:20:06.777050 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" Feb 20 00:20:07 crc kubenswrapper[4750]: I0220 00:20:07.215642 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-lg2cd"] Feb 20 00:20:07 crc kubenswrapper[4750]: W0220 00:20:07.223656 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eeff3f9_8995_450d_af18_d129c823ec09.slice/crio-d3d453376a32157ea0bbd8454cec7ddfdd60000caebdf25b1037496471d10c1c WatchSource:0}: Error finding container d3d453376a32157ea0bbd8454cec7ddfdd60000caebdf25b1037496471d10c1c: Status 404 returned error can't find the container with id d3d453376a32157ea0bbd8454cec7ddfdd60000caebdf25b1037496471d10c1c Feb 20 00:20:07 crc kubenswrapper[4750]: I0220 00:20:07.609742 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" event={"ID":"1eeff3f9-8995-450d-af18-d129c823ec09","Type":"ContainerStarted","Data":"d3d453376a32157ea0bbd8454cec7ddfdd60000caebdf25b1037496471d10c1c"} Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.736056 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.737610 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.739631 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.740008 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.740064 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-gsktv" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.740368 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.754024 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.880261 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-td2k7"] Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.881170 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.884461 4750 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-ccrr7" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.895167 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-td2k7"] Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.895979 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896020 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896041 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896077 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896098 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896142 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896170 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896191 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896222 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896242 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896291 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5zqh\" (UniqueName: \"kubernetes.io/projected/03210243-d15e-4b75-b85c-e4c6efb10fc9-kube-api-access-t5zqh\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.896312 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.997910 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.997952 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.997972 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998018 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998040 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998058 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998077 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998101 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998153 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998180 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/34841f08-32f2-4738-8241-7b5c9e0408d7-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-td2k7\" (UID: \"34841f08-32f2-4738-8241-7b5c9e0408d7\") " pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998202 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998249 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5zqh\" (UniqueName: \"kubernetes.io/projected/03210243-d15e-4b75-b85c-e4c6efb10fc9-kube-api-access-t5zqh\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998271 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw9kj\" (UniqueName: \"kubernetes.io/projected/34841f08-32f2-4738-8241-7b5c9e0408d7-kube-api-access-nw9kj\") pod \"cert-manager-webhook-6888856db4-td2k7\" (UID: \"34841f08-32f2-4738-8241-7b5c9e0408d7\") " pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998292 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998461 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998515 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.998946 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.999215 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.999438 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.999627 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.999695 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:08 crc kubenswrapper[4750]: I0220 00:20:08.999843 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.000084 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.003019 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.014585 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.031312 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5zqh\" (UniqueName: \"kubernetes.io/projected/03210243-d15e-4b75-b85c-e4c6efb10fc9-kube-api-access-t5zqh\") pod \"service-telemetry-operator-1-build\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.053251 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.099839 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/34841f08-32f2-4738-8241-7b5c9e0408d7-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-td2k7\" (UID: \"34841f08-32f2-4738-8241-7b5c9e0408d7\") " pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.099920 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw9kj\" (UniqueName: \"kubernetes.io/projected/34841f08-32f2-4738-8241-7b5c9e0408d7-kube-api-access-nw9kj\") pod \"cert-manager-webhook-6888856db4-td2k7\" (UID: \"34841f08-32f2-4738-8241-7b5c9e0408d7\") " pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.115519 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw9kj\" (UniqueName: \"kubernetes.io/projected/34841f08-32f2-4738-8241-7b5c9e0408d7-kube-api-access-nw9kj\") pod \"cert-manager-webhook-6888856db4-td2k7\" (UID: \"34841f08-32f2-4738-8241-7b5c9e0408d7\") " pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.115597 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/34841f08-32f2-4738-8241-7b5c9e0408d7-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-td2k7\" (UID: \"34841f08-32f2-4738-8241-7b5c9e0408d7\") " pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.196170 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.541076 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.596333 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-td2k7"] Feb 20 00:20:09 crc kubenswrapper[4750]: W0220 00:20:09.600534 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34841f08_32f2_4738_8241_7b5c9e0408d7.slice/crio-9f970b32970ac08b466fa6c0400c1dfe490d5ab42a1da83a0ad1d087abd5f5e0 WatchSource:0}: Error finding container 9f970b32970ac08b466fa6c0400c1dfe490d5ab42a1da83a0ad1d087abd5f5e0: Status 404 returned error can't find the container with id 9f970b32970ac08b466fa6c0400c1dfe490d5ab42a1da83a0ad1d087abd5f5e0 Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.624945 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" event={"ID":"34841f08-32f2-4738-8241-7b5c9e0408d7","Type":"ContainerStarted","Data":"9f970b32970ac08b466fa6c0400c1dfe490d5ab42a1da83a0ad1d087abd5f5e0"} Feb 20 00:20:09 crc kubenswrapper[4750]: I0220 00:20:09.625846 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"03210243-d15e-4b75-b85c-e4c6efb10fc9","Type":"ContainerStarted","Data":"b0d4e89aaa077446aa7ef06b70be7be2940fb3b4ce497ed951d0a0a74a2f32d4"} Feb 20 00:20:12 crc kubenswrapper[4750]: I0220 00:20:12.649493 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" event={"ID":"1eeff3f9-8995-450d-af18-d129c823ec09","Type":"ContainerStarted","Data":"3bc044dd17e672be9850ea67d5c3a34eb54b8360dcba0ad60bdafdbcc58f41e7"} Feb 20 00:20:12 crc kubenswrapper[4750]: I0220 00:20:12.652963 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" event={"ID":"34841f08-32f2-4738-8241-7b5c9e0408d7","Type":"ContainerStarted","Data":"8bafb4b2315998207658fd1bb11e5a44a56bdb5848d4df3ff8e3051093f7c022"} Feb 20 00:20:12 crc kubenswrapper[4750]: I0220 00:20:12.653129 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:12 crc kubenswrapper[4750]: I0220 00:20:12.665365 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-lg2cd" podStartSLOduration=2.022658257 podStartE2EDuration="6.665345268s" podCreationTimestamp="2026-02-20 00:20:06 +0000 UTC" firstStartedPulling="2026-02-20 00:20:07.226089435 +0000 UTC m=+691.420925494" lastFinishedPulling="2026-02-20 00:20:11.868776456 +0000 UTC m=+696.063612505" observedRunningTime="2026-02-20 00:20:12.661940872 +0000 UTC m=+696.856776931" watchObservedRunningTime="2026-02-20 00:20:12.665345268 +0000 UTC m=+696.860181317" Feb 20 00:20:12 crc kubenswrapper[4750]: I0220 00:20:12.684817 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" podStartSLOduration=2.421344154 podStartE2EDuration="4.684796042s" podCreationTimestamp="2026-02-20 00:20:08 +0000 UTC" firstStartedPulling="2026-02-20 00:20:09.602747303 +0000 UTC m=+693.797583352" lastFinishedPulling="2026-02-20 00:20:11.866199191 +0000 UTC m=+696.061035240" observedRunningTime="2026-02-20 00:20:12.682293458 +0000 UTC m=+696.877129527" watchObservedRunningTime="2026-02-20 00:20:12.684796042 +0000 UTC m=+696.879632101" Feb 20 00:20:15 crc kubenswrapper[4750]: I0220 00:20:15.682996 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"03210243-d15e-4b75-b85c-e4c6efb10fc9","Type":"ContainerStarted","Data":"7aeec7e1bfcc8a7a4af5f21bd8c7ce38fd60be6ed6c8fd21685f15159a0cc7e7"} Feb 20 00:20:15 crc kubenswrapper[4750]: I0220 00:20:15.685011 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c82277c8-3f73-4244-a092-0117400628a8","Type":"ContainerStarted","Data":"7fd3b3b01b3e610f0d9929dbcb43ae9e63c033bb05b2da89e926df95ba8fd959"} Feb 20 00:20:16 crc kubenswrapper[4750]: I0220 00:20:16.694335 4750 generic.go:334] "Generic (PLEG): container finished" podID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerID="7aeec7e1bfcc8a7a4af5f21bd8c7ce38fd60be6ed6c8fd21685f15159a0cc7e7" exitCode=0 Feb 20 00:20:16 crc kubenswrapper[4750]: I0220 00:20:16.694409 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"03210243-d15e-4b75-b85c-e4c6efb10fc9","Type":"ContainerDied","Data":"7aeec7e1bfcc8a7a4af5f21bd8c7ce38fd60be6ed6c8fd21685f15159a0cc7e7"} Feb 20 00:20:17 crc kubenswrapper[4750]: I0220 00:20:17.703193 4750 generic.go:334] "Generic (PLEG): container finished" podID="c82277c8-3f73-4244-a092-0117400628a8" containerID="7fd3b3b01b3e610f0d9929dbcb43ae9e63c033bb05b2da89e926df95ba8fd959" exitCode=0 Feb 20 00:20:17 crc kubenswrapper[4750]: I0220 00:20:17.703329 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c82277c8-3f73-4244-a092-0117400628a8","Type":"ContainerDied","Data":"7fd3b3b01b3e610f0d9929dbcb43ae9e63c033bb05b2da89e926df95ba8fd959"} Feb 20 00:20:18 crc kubenswrapper[4750]: I0220 00:20:18.740658 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 20 00:20:19 crc kubenswrapper[4750]: I0220 00:20:19.199693 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-td2k7" Feb 20 00:20:19 crc kubenswrapper[4750]: I0220 00:20:19.719186 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"03210243-d15e-4b75-b85c-e4c6efb10fc9","Type":"ContainerStarted","Data":"8caa90ac141fa6b129dc62bb3e8bd2c1c4e12ec65945b0ccf3e692f1a61b3c22"} Feb 20 00:20:19 crc kubenswrapper[4750]: I0220 00:20:19.721530 4750 generic.go:334] "Generic (PLEG): container finished" podID="c82277c8-3f73-4244-a092-0117400628a8" containerID="3fcbf81e7be88405dae8d82df72ca967865400a84cbc2e8f2b9bc7499b1b1718" exitCode=0 Feb 20 00:20:19 crc kubenswrapper[4750]: I0220 00:20:19.721578 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c82277c8-3f73-4244-a092-0117400628a8","Type":"ContainerDied","Data":"3fcbf81e7be88405dae8d82df72ca967865400a84cbc2e8f2b9bc7499b1b1718"} Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.340301 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.341338 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.345164 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.345338 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.345151 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.364251 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.508997 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509049 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509176 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509211 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509231 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509321 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sc8x\" (UniqueName: \"kubernetes.io/projected/2222accf-d7d5-4210-8431-b321f9be2625-kube-api-access-5sc8x\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509358 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509387 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509418 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509443 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509548 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.509605 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610320 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610386 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sc8x\" (UniqueName: \"kubernetes.io/projected/2222accf-d7d5-4210-8431-b321f9be2625-kube-api-access-5sc8x\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610416 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610440 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610460 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610478 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610498 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610522 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610564 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610583 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610601 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.610620 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.611022 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.611157 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.611212 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.611284 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.611804 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.611804 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.612062 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.613381 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.614234 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.620461 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.622629 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.643593 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sc8x\" (UniqueName: \"kubernetes.io/projected/2222accf-d7d5-4210-8431-b321f9be2625-kube-api-access-5sc8x\") pod \"service-telemetry-operator-2-build\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.711152 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.729412 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerName="docker-build" containerID="cri-o://8caa90ac141fa6b129dc62bb3e8bd2c1c4e12ec65945b0ccf3e692f1a61b3c22" gracePeriod=30 Feb 20 00:20:20 crc kubenswrapper[4750]: I0220 00:20:20.753299 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-1-build" podStartSLOduration=7.082814781 podStartE2EDuration="12.753280464s" podCreationTimestamp="2026-02-20 00:20:08 +0000 UTC" firstStartedPulling="2026-02-20 00:20:09.560588355 +0000 UTC m=+693.755424404" lastFinishedPulling="2026-02-20 00:20:15.231053998 +0000 UTC m=+699.425890087" observedRunningTime="2026-02-20 00:20:20.75194547 +0000 UTC m=+704.946781539" watchObservedRunningTime="2026-02-20 00:20:20.753280464 +0000 UTC m=+704.948116523" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.598749 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-pd5km"] Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.600623 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.602220 4750 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-pb9x6" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.604858 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-pd5km"] Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.627829 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsfxx\" (UniqueName: \"kubernetes.io/projected/16f80a94-8f72-4baf-944f-8934f7a2cdc5-kube-api-access-fsfxx\") pod \"cert-manager-545d4d4674-pd5km\" (UID: \"16f80a94-8f72-4baf-944f-8934f7a2cdc5\") " pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.627888 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16f80a94-8f72-4baf-944f-8934f7a2cdc5-bound-sa-token\") pod \"cert-manager-545d4d4674-pd5km\" (UID: \"16f80a94-8f72-4baf-944f-8934f7a2cdc5\") " pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.729661 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsfxx\" (UniqueName: \"kubernetes.io/projected/16f80a94-8f72-4baf-944f-8934f7a2cdc5-kube-api-access-fsfxx\") pod \"cert-manager-545d4d4674-pd5km\" (UID: \"16f80a94-8f72-4baf-944f-8934f7a2cdc5\") " pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.729753 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16f80a94-8f72-4baf-944f-8934f7a2cdc5-bound-sa-token\") pod \"cert-manager-545d4d4674-pd5km\" (UID: \"16f80a94-8f72-4baf-944f-8934f7a2cdc5\") " pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.757756 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsfxx\" (UniqueName: \"kubernetes.io/projected/16f80a94-8f72-4baf-944f-8934f7a2cdc5-kube-api-access-fsfxx\") pod \"cert-manager-545d4d4674-pd5km\" (UID: \"16f80a94-8f72-4baf-944f-8934f7a2cdc5\") " pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.762303 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16f80a94-8f72-4baf-944f-8934f7a2cdc5-bound-sa-token\") pod \"cert-manager-545d4d4674-pd5km\" (UID: \"16f80a94-8f72-4baf-944f-8934f7a2cdc5\") " pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:21 crc kubenswrapper[4750]: I0220 00:20:21.945464 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-pd5km" Feb 20 00:20:22 crc kubenswrapper[4750]: I0220 00:20:22.725512 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 20 00:20:22 crc kubenswrapper[4750]: I0220 00:20:22.753381 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"2222accf-d7d5-4210-8431-b321f9be2625","Type":"ContainerStarted","Data":"6efc927c607aa516d09c13a5ba271afc5a2ac2f6353917767b38a8d8e78dcac7"} Feb 20 00:20:22 crc kubenswrapper[4750]: I0220 00:20:22.797754 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-pd5km"] Feb 20 00:20:23 crc kubenswrapper[4750]: I0220 00:20:23.763525 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-pd5km" event={"ID":"16f80a94-8f72-4baf-944f-8934f7a2cdc5","Type":"ContainerStarted","Data":"9355675077c7290e8f7127296cb40524177763f4a292161819ec6860304b4792"} Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.776769 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"c82277c8-3f73-4244-a092-0117400628a8","Type":"ContainerStarted","Data":"0cf6905f34e4aec19f2436810a9a00dc03469500d781e36133028f389ecfd4c5"} Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.777308 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.779746 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"2222accf-d7d5-4210-8431-b321f9be2625","Type":"ContainerStarted","Data":"ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100"} Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.781841 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-pd5km" event={"ID":"16f80a94-8f72-4baf-944f-8934f7a2cdc5","Type":"ContainerStarted","Data":"bc0670e46cd77f082711b0b52b16bc42d710e1e2e51fc4a6bfdc820552643f57"} Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.784411 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_03210243-d15e-4b75-b85c-e4c6efb10fc9/docker-build/0.log" Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.784862 4750 generic.go:334] "Generic (PLEG): container finished" podID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerID="8caa90ac141fa6b129dc62bb3e8bd2c1c4e12ec65945b0ccf3e692f1a61b3c22" exitCode=1 Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.784914 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"03210243-d15e-4b75-b85c-e4c6efb10fc9","Type":"ContainerDied","Data":"8caa90ac141fa6b129dc62bb3e8bd2c1c4e12ec65945b0ccf3e692f1a61b3c22"} Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.857354 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=10.800213627 podStartE2EDuration="53.857331075s" podCreationTimestamp="2026-02-20 00:19:31 +0000 UTC" firstStartedPulling="2026-02-20 00:19:32.173226312 +0000 UTC m=+656.368062361" lastFinishedPulling="2026-02-20 00:20:15.23034375 +0000 UTC m=+699.425179809" observedRunningTime="2026-02-20 00:20:24.848330511 +0000 UTC m=+709.043166590" watchObservedRunningTime="2026-02-20 00:20:24.857331075 +0000 UTC m=+709.052167124" Feb 20 00:20:24 crc kubenswrapper[4750]: I0220 00:20:24.912912 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-pd5km" podStartSLOduration=3.912895578 podStartE2EDuration="3.912895578s" podCreationTimestamp="2026-02-20 00:20:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 00:20:24.912744444 +0000 UTC m=+709.107580503" watchObservedRunningTime="2026-02-20 00:20:24.912895578 +0000 UTC m=+709.107731627" Feb 20 00:20:25 crc kubenswrapper[4750]: E0220 00:20:25.001597 4750 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=2779325518433787891, SKID=, AKID=51:D2:4F:9F:30:C7:B6:3D:15:1B:65:4B:CC:7E:11:9A:B8:75:96:5A failed: x509: certificate signed by unknown authority" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.375648 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_03210243-d15e-4b75-b85c-e4c6efb10fc9/docker-build/0.log" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.376033 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.509874 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-push\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510254 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-proxy-ca-bundles\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510286 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-pull\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510313 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-system-configs\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510346 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-root\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510369 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-blob-cache\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510391 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-ca-bundles\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510408 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5zqh\" (UniqueName: \"kubernetes.io/projected/03210243-d15e-4b75-b85c-e4c6efb10fc9-kube-api-access-t5zqh\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510435 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-node-pullsecrets\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510462 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildworkdir\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510481 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-run\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510504 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildcachedir\") pod \"03210243-d15e-4b75-b85c-e4c6efb10fc9\" (UID: \"03210243-d15e-4b75-b85c-e4c6efb10fc9\") " Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510556 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510692 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510740 4750 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.510859 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.511003 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.511075 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.511103 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.511367 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.511392 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.512270 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.526021 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-push" (OuterVolumeSpecName: "builder-dockercfg-gsktv-push") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "builder-dockercfg-gsktv-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.526092 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03210243-d15e-4b75-b85c-e4c6efb10fc9-kube-api-access-t5zqh" (OuterVolumeSpecName: "kube-api-access-t5zqh") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "kube-api-access-t5zqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.526130 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-pull" (OuterVolumeSpecName: "builder-dockercfg-gsktv-pull") pod "03210243-d15e-4b75-b85c-e4c6efb10fc9" (UID: "03210243-d15e-4b75-b85c-e4c6efb10fc9"). InnerVolumeSpecName "builder-dockercfg-gsktv-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.612259 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-push\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.612801 4750 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.612957 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/03210243-d15e-4b75-b85c-e4c6efb10fc9-builder-dockercfg-gsktv-pull\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.613085 4750 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.613248 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.613368 4750 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.613521 4750 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03210243-d15e-4b75-b85c-e4c6efb10fc9-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.613653 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5zqh\" (UniqueName: \"kubernetes.io/projected/03210243-d15e-4b75-b85c-e4c6efb10fc9-kube-api-access-t5zqh\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.613773 4750 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.613904 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/03210243-d15e-4b75-b85c-e4c6efb10fc9-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.614022 4750 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/03210243-d15e-4b75-b85c-e4c6efb10fc9-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.793998 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_03210243-d15e-4b75-b85c-e4c6efb10fc9/docker-build/0.log" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.794520 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"03210243-d15e-4b75-b85c-e4c6efb10fc9","Type":"ContainerDied","Data":"b0d4e89aaa077446aa7ef06b70be7be2940fb3b4ce497ed951d0a0a74a2f32d4"} Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.794601 4750 scope.go:117] "RemoveContainer" containerID="8caa90ac141fa6b129dc62bb3e8bd2c1c4e12ec65945b0ccf3e692f1a61b3c22" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.795587 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.813571 4750 scope.go:117] "RemoveContainer" containerID="7aeec7e1bfcc8a7a4af5f21bd8c7ce38fd60be6ed6c8fd21685f15159a0cc7e7" Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.865620 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 20 00:20:25 crc kubenswrapper[4750]: I0220 00:20:25.884028 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 20 00:20:26 crc kubenswrapper[4750]: I0220 00:20:26.035823 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 20 00:20:26 crc kubenswrapper[4750]: I0220 00:20:26.568903 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03210243-d15e-4b75-b85c-e4c6efb10fc9" path="/var/lib/kubelet/pods/03210243-d15e-4b75-b85c-e4c6efb10fc9/volumes" Feb 20 00:20:26 crc kubenswrapper[4750]: I0220 00:20:26.802197 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-2-build" podUID="2222accf-d7d5-4210-8431-b321f9be2625" containerName="git-clone" containerID="cri-o://ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100" gracePeriod=30 Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.183793 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_2222accf-d7d5-4210-8431-b321f9be2625/git-clone/0.log" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.183951 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338710 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-node-pullsecrets\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338767 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-system-configs\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338787 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-buildworkdir\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338827 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-pull\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338851 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-ca-bundles\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338860 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338887 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-proxy-ca-bundles\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338907 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-push\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338929 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-build-blob-cache\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338956 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-root\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.338988 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sc8x\" (UniqueName: \"kubernetes.io/projected/2222accf-d7d5-4210-8431-b321f9be2625-kube-api-access-5sc8x\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.339010 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-buildcachedir\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.339043 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-run\") pod \"2222accf-d7d5-4210-8431-b321f9be2625\" (UID: \"2222accf-d7d5-4210-8431-b321f9be2625\") " Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.339270 4750 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.339558 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.339654 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.339799 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.339888 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.340209 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.340234 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.340432 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.341047 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.345835 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-push" (OuterVolumeSpecName: "builder-dockercfg-gsktv-push") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "builder-dockercfg-gsktv-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.346310 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-pull" (OuterVolumeSpecName: "builder-dockercfg-gsktv-pull") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "builder-dockercfg-gsktv-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.346834 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2222accf-d7d5-4210-8431-b321f9be2625-kube-api-access-5sc8x" (OuterVolumeSpecName: "kube-api-access-5sc8x") pod "2222accf-d7d5-4210-8431-b321f9be2625" (UID: "2222accf-d7d5-4210-8431-b321f9be2625"). InnerVolumeSpecName "kube-api-access-5sc8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441043 4750 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/2222accf-d7d5-4210-8431-b321f9be2625-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441085 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441101 4750 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441114 4750 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441131 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-pull\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441160 4750 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441176 4750 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2222accf-d7d5-4210-8431-b321f9be2625-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441189 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/2222accf-d7d5-4210-8431-b321f9be2625-builder-dockercfg-gsktv-push\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441203 4750 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441215 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/2222accf-d7d5-4210-8431-b321f9be2625-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.441229 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sc8x\" (UniqueName: \"kubernetes.io/projected/2222accf-d7d5-4210-8431-b321f9be2625-kube-api-access-5sc8x\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.808976 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_2222accf-d7d5-4210-8431-b321f9be2625/git-clone/0.log" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.809046 4750 generic.go:334] "Generic (PLEG): container finished" podID="2222accf-d7d5-4210-8431-b321f9be2625" containerID="ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100" exitCode=1 Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.809087 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"2222accf-d7d5-4210-8431-b321f9be2625","Type":"ContainerDied","Data":"ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100"} Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.809140 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.809132 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"2222accf-d7d5-4210-8431-b321f9be2625","Type":"ContainerDied","Data":"6efc927c607aa516d09c13a5ba271afc5a2ac2f6353917767b38a8d8e78dcac7"} Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.809184 4750 scope.go:117] "RemoveContainer" containerID="ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.829244 4750 scope.go:117] "RemoveContainer" containerID="ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100" Feb 20 00:20:27 crc kubenswrapper[4750]: E0220 00:20:27.830150 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100\": container with ID starting with ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100 not found: ID does not exist" containerID="ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.830229 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100"} err="failed to get container status \"ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100\": rpc error: code = NotFound desc = could not find container \"ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100\": container with ID starting with ffd81d5a0c624d91d353bb02a6d8d506f4e01e8039a9e24a2d69cd283a1f9100 not found: ID does not exist" Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.867979 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 20 00:20:27 crc kubenswrapper[4750]: I0220 00:20:27.872473 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 20 00:20:28 crc kubenswrapper[4750]: I0220 00:20:28.573917 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2222accf-d7d5-4210-8431-b321f9be2625" path="/var/lib/kubelet/pods/2222accf-d7d5-4210-8431-b321f9be2625/volumes" Feb 20 00:20:36 crc kubenswrapper[4750]: I0220 00:20:36.831620 4750 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="c82277c8-3f73-4244-a092-0117400628a8" containerName="elasticsearch" probeResult="failure" output=< Feb 20 00:20:36 crc kubenswrapper[4750]: {"timestamp": "2026-02-20T00:20:36+00:00", "message": "readiness probe failed", "curl_rc": "7"} Feb 20 00:20:36 crc kubenswrapper[4750]: > Feb 20 00:20:36 crc kubenswrapper[4750]: I0220 00:20:36.973428 4750 scope.go:117] "RemoveContainer" containerID="a2146a4334faa3fb46dc02bb85b8b98702ef1196aa4d3ab4db4a283a987f8ef2" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.025252 4750 scope.go:117] "RemoveContainer" containerID="fcaffaf887b0a111113bca54d74f736b947d57f61f40f79ae53729ac53914c6a" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.077611 4750 scope.go:117] "RemoveContainer" containerID="84e7289668ac5c08c2fc2613e3fbcd8348cd606db6d0f5b844659e317311cb09" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.448016 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 20 00:20:37 crc kubenswrapper[4750]: E0220 00:20:37.448321 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerName="docker-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.448348 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerName="docker-build" Feb 20 00:20:37 crc kubenswrapper[4750]: E0220 00:20:37.448371 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2222accf-d7d5-4210-8431-b321f9be2625" containerName="git-clone" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.448380 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="2222accf-d7d5-4210-8431-b321f9be2625" containerName="git-clone" Feb 20 00:20:37 crc kubenswrapper[4750]: E0220 00:20:37.448402 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerName="manage-dockerfile" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.448410 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerName="manage-dockerfile" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.448540 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="03210243-d15e-4b75-b85c-e4c6efb10fc9" containerName="docker-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.448555 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="2222accf-d7d5-4210-8431-b321f9be2625" containerName="git-clone" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.449563 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.451165 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-gsktv" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.451456 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-ca" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.451743 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-sys-config" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.452891 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-global-ca" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.475991 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575070 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575425 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575462 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575485 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575530 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575564 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575597 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575618 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575643 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575682 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbqfn\" (UniqueName: \"kubernetes.io/projected/8a946bc3-32b8-4246-8f62-985846c0f1c4-kube-api-access-bbqfn\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575713 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.575738 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.677318 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.677533 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.677605 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.677701 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.677782 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.677850 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.677933 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.678009 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.678097 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.678185 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.678263 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.678350 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbqfn\" (UniqueName: \"kubernetes.io/projected/8a946bc3-32b8-4246-8f62-985846c0f1c4-kube-api-access-bbqfn\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.679022 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.679154 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.679377 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.679599 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.679832 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.680485 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.680665 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.681001 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.681298 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.686652 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.687042 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.697005 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbqfn\" (UniqueName: \"kubernetes.io/projected/8a946bc3-32b8-4246-8f62-985846c0f1c4-kube-api-access-bbqfn\") pod \"service-telemetry-operator-3-build\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:37 crc kubenswrapper[4750]: I0220 00:20:37.768235 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:38 crc kubenswrapper[4750]: I0220 00:20:38.219727 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 20 00:20:38 crc kubenswrapper[4750]: I0220 00:20:38.902269 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"8a946bc3-32b8-4246-8f62-985846c0f1c4","Type":"ContainerStarted","Data":"e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04"} Feb 20 00:20:38 crc kubenswrapper[4750]: I0220 00:20:38.902658 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"8a946bc3-32b8-4246-8f62-985846c0f1c4","Type":"ContainerStarted","Data":"53b97e530a36599bda7aa473b3b45cd15db604a39ff6cd728436e388fd2eb758"} Feb 20 00:20:38 crc kubenswrapper[4750]: E0220 00:20:38.978578 4750 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=2779325518433787891, SKID=, AKID=51:D2:4F:9F:30:C7:B6:3D:15:1B:65:4B:CC:7E:11:9A:B8:75:96:5A failed: x509: certificate signed by unknown authority" Feb 20 00:20:40 crc kubenswrapper[4750]: I0220 00:20:40.004263 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 20 00:20:40 crc kubenswrapper[4750]: I0220 00:20:40.917293 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-3-build" podUID="8a946bc3-32b8-4246-8f62-985846c0f1c4" containerName="git-clone" containerID="cri-o://e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04" gracePeriod=30 Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.315894 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_8a946bc3-32b8-4246-8f62-985846c0f1c4/git-clone/0.log" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.316291 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445424 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-proxy-ca-bundles\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445508 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-node-pullsecrets\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445550 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-run\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445583 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildworkdir\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445613 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-root\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445630 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-blob-cache\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445649 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-push\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445668 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbqfn\" (UniqueName: \"kubernetes.io/projected/8a946bc3-32b8-4246-8f62-985846c0f1c4-kube-api-access-bbqfn\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445712 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-ca-bundles\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445728 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-pull\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445750 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildcachedir\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.445767 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-system-configs\") pod \"8a946bc3-32b8-4246-8f62-985846c0f1c4\" (UID: \"8a946bc3-32b8-4246-8f62-985846c0f1c4\") " Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.446426 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.446464 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.446589 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.446652 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.446759 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.447112 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.447810 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.447905 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.448151 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.452312 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-push" (OuterVolumeSpecName: "builder-dockercfg-gsktv-push") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "builder-dockercfg-gsktv-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.452546 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a946bc3-32b8-4246-8f62-985846c0f1c4-kube-api-access-bbqfn" (OuterVolumeSpecName: "kube-api-access-bbqfn") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "kube-api-access-bbqfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.453616 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-pull" (OuterVolumeSpecName: "builder-dockercfg-gsktv-pull") pod "8a946bc3-32b8-4246-8f62-985846c0f1c4" (UID: "8a946bc3-32b8-4246-8f62-985846c0f1c4"). InnerVolumeSpecName "builder-dockercfg-gsktv-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546723 4750 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546768 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546784 4750 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546797 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-push\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546809 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbqfn\" (UniqueName: \"kubernetes.io/projected/8a946bc3-32b8-4246-8f62-985846c0f1c4-kube-api-access-bbqfn\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546822 4750 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546833 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/8a946bc3-32b8-4246-8f62-985846c0f1c4-builder-dockercfg-gsktv-pull\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546845 4750 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546856 4750 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546866 4750 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a946bc3-32b8-4246-8f62-985846c0f1c4-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546883 4750 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8a946bc3-32b8-4246-8f62-985846c0f1c4-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.546895 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8a946bc3-32b8-4246-8f62-985846c0f1c4-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.927892 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_8a946bc3-32b8-4246-8f62-985846c0f1c4/git-clone/0.log" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.928621 4750 generic.go:334] "Generic (PLEG): container finished" podID="8a946bc3-32b8-4246-8f62-985846c0f1c4" containerID="e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04" exitCode=1 Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.928821 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"8a946bc3-32b8-4246-8f62-985846c0f1c4","Type":"ContainerDied","Data":"e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04"} Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.929365 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"8a946bc3-32b8-4246-8f62-985846c0f1c4","Type":"ContainerDied","Data":"53b97e530a36599bda7aa473b3b45cd15db604a39ff6cd728436e388fd2eb758"} Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.929541 4750 scope.go:117] "RemoveContainer" containerID="e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.928935 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.966822 4750 scope.go:117] "RemoveContainer" containerID="e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04" Feb 20 00:20:41 crc kubenswrapper[4750]: E0220 00:20:41.967268 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04\": container with ID starting with e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04 not found: ID does not exist" containerID="e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.967313 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04"} err="failed to get container status \"e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04\": rpc error: code = NotFound desc = could not find container \"e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04\": container with ID starting with e2787e46cc82c05ec74864b2f369374ff36a6f9fe3f1da03af28002e0233ba04 not found: ID does not exist" Feb 20 00:20:41 crc kubenswrapper[4750]: I0220 00:20:41.991537 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 20 00:20:42 crc kubenswrapper[4750]: I0220 00:20:42.001621 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 20 00:20:42 crc kubenswrapper[4750]: I0220 00:20:42.335656 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Feb 20 00:20:42 crc kubenswrapper[4750]: I0220 00:20:42.567829 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a946bc3-32b8-4246-8f62-985846c0f1c4" path="/var/lib/kubelet/pods/8a946bc3-32b8-4246-8f62-985846c0f1c4/volumes" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.489682 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 20 00:20:51 crc kubenswrapper[4750]: E0220 00:20:51.490576 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a946bc3-32b8-4246-8f62-985846c0f1c4" containerName="git-clone" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.490598 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a946bc3-32b8-4246-8f62-985846c0f1c4" containerName="git-clone" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.490748 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a946bc3-32b8-4246-8f62-985846c0f1c4" containerName="git-clone" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.491830 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.494584 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-global-ca" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.494607 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-sys-config" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.494934 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-ca" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.494971 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-gsktv" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.520953 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586275 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586314 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586342 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586359 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586381 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586416 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzv7m\" (UniqueName: \"kubernetes.io/projected/030f472b-1455-4b0e-ac5f-ebb615012487-kube-api-access-fzv7m\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586470 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586521 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586544 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586585 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586617 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.586632 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687466 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687589 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687654 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687708 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687772 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzv7m\" (UniqueName: \"kubernetes.io/projected/030f472b-1455-4b0e-ac5f-ebb615012487-kube-api-access-fzv7m\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687858 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687898 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.687924 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688016 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688101 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688298 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688356 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688408 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688442 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688608 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688672 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.688742 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.689610 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.689962 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.690093 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.690733 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.695404 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.699734 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.733847 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzv7m\" (UniqueName: \"kubernetes.io/projected/030f472b-1455-4b0e-ac5f-ebb615012487-kube-api-access-fzv7m\") pod \"service-telemetry-operator-4-build\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:51 crc kubenswrapper[4750]: I0220 00:20:51.812302 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:52 crc kubenswrapper[4750]: I0220 00:20:52.075910 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 20 00:20:53 crc kubenswrapper[4750]: I0220 00:20:53.007060 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"030f472b-1455-4b0e-ac5f-ebb615012487","Type":"ContainerStarted","Data":"30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767"} Feb 20 00:20:53 crc kubenswrapper[4750]: I0220 00:20:53.007555 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"030f472b-1455-4b0e-ac5f-ebb615012487","Type":"ContainerStarted","Data":"f0d18ac9ec3e6675327c7389cbf5be1e943c5eb1291d01ccabf42ae9c1af4014"} Feb 20 00:20:53 crc kubenswrapper[4750]: E0220 00:20:53.085858 4750 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=2779325518433787891, SKID=, AKID=51:D2:4F:9F:30:C7:B6:3D:15:1B:65:4B:CC:7E:11:9A:B8:75:96:5A failed: x509: certificate signed by unknown authority" Feb 20 00:20:54 crc kubenswrapper[4750]: I0220 00:20:54.116078 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.026147 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-4-build" podUID="030f472b-1455-4b0e-ac5f-ebb615012487" containerName="git-clone" containerID="cri-o://30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767" gracePeriod=30 Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.491374 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_030f472b-1455-4b0e-ac5f-ebb615012487/git-clone/0.log" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.491724 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.642868 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-build-blob-cache\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.642966 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-buildworkdir\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643023 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-root\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643092 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-proxy-ca-bundles\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643223 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-node-pullsecrets\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643259 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-system-configs\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643303 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-ca-bundles\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643374 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643537 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643707 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643910 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.643994 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-run\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644039 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzv7m\" (UniqueName: \"kubernetes.io/projected/030f472b-1455-4b0e-ac5f-ebb615012487-kube-api-access-fzv7m\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644091 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-pull\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644174 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-buildcachedir\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644245 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-push\") pod \"030f472b-1455-4b0e-ac5f-ebb615012487\" (UID: \"030f472b-1455-4b0e-ac5f-ebb615012487\") " Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644328 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644559 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644554 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644597 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644841 4750 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644877 4750 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644901 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644927 4750 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644953 4750 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.644977 4750 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.645000 4750 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/030f472b-1455-4b0e-ac5f-ebb615012487-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.645023 4750 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/030f472b-1455-4b0e-ac5f-ebb615012487-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.646666 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.653897 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-pull" (OuterVolumeSpecName: "builder-dockercfg-gsktv-pull") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "builder-dockercfg-gsktv-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.653933 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-push" (OuterVolumeSpecName: "builder-dockercfg-gsktv-push") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "builder-dockercfg-gsktv-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.655786 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/030f472b-1455-4b0e-ac5f-ebb615012487-kube-api-access-fzv7m" (OuterVolumeSpecName: "kube-api-access-fzv7m") pod "030f472b-1455-4b0e-ac5f-ebb615012487" (UID: "030f472b-1455-4b0e-ac5f-ebb615012487"). InnerVolumeSpecName "kube-api-access-fzv7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.746618 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-pull\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.746677 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/030f472b-1455-4b0e-ac5f-ebb615012487-builder-dockercfg-gsktv-push\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.746695 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/030f472b-1455-4b0e-ac5f-ebb615012487-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:55 crc kubenswrapper[4750]: I0220 00:20:55.746712 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzv7m\" (UniqueName: \"kubernetes.io/projected/030f472b-1455-4b0e-ac5f-ebb615012487-kube-api-access-fzv7m\") on node \"crc\" DevicePath \"\"" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.036483 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_030f472b-1455-4b0e-ac5f-ebb615012487/git-clone/0.log" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.036543 4750 generic.go:334] "Generic (PLEG): container finished" podID="030f472b-1455-4b0e-ac5f-ebb615012487" containerID="30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767" exitCode=1 Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.036575 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"030f472b-1455-4b0e-ac5f-ebb615012487","Type":"ContainerDied","Data":"30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767"} Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.036605 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"030f472b-1455-4b0e-ac5f-ebb615012487","Type":"ContainerDied","Data":"f0d18ac9ec3e6675327c7389cbf5be1e943c5eb1291d01ccabf42ae9c1af4014"} Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.036625 4750 scope.go:117] "RemoveContainer" containerID="30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.036662 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.066874 4750 scope.go:117] "RemoveContainer" containerID="30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767" Feb 20 00:20:56 crc kubenswrapper[4750]: E0220 00:20:56.067453 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767\": container with ID starting with 30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767 not found: ID does not exist" containerID="30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.067600 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767"} err="failed to get container status \"30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767\": rpc error: code = NotFound desc = could not find container \"30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767\": container with ID starting with 30ad7e23745d26478b143773895396d2ef0d78698548b847600c5411cb34a767 not found: ID does not exist" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.091555 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.100870 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 20 00:20:56 crc kubenswrapper[4750]: E0220 00:20:56.167267 4750 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod030f472b_1455_4b0e_ac5f_ebb615012487.slice/crio-f0d18ac9ec3e6675327c7389cbf5be1e943c5eb1291d01ccabf42ae9c1af4014\": RecentStats: unable to find data in memory cache]" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.570357 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="030f472b-1455-4b0e-ac5f-ebb615012487" path="/var/lib/kubelet/pods/030f472b-1455-4b0e-ac5f-ebb615012487/volumes" Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.669958 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:20:56 crc kubenswrapper[4750]: I0220 00:20:56.670066 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.096438 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 20 00:21:06 crc kubenswrapper[4750]: E0220 00:21:06.097646 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="030f472b-1455-4b0e-ac5f-ebb615012487" containerName="git-clone" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.097677 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="030f472b-1455-4b0e-ac5f-ebb615012487" containerName="git-clone" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.097976 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="030f472b-1455-4b0e-ac5f-ebb615012487" containerName="git-clone" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.100099 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.107379 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-sys-config" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.107801 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-ca" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.108346 4750 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-gsktv" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.108493 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-global-ca" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.128014 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144450 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144508 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144539 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144560 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144596 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144629 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144651 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144688 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f66fm\" (UniqueName: \"kubernetes.io/projected/079821a6-579a-43df-bfbb-394c998cfac8-kube-api-access-f66fm\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144718 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144743 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144766 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.144804 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.245862 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.245937 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.245967 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.245988 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246029 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246061 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246073 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246080 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246179 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f66fm\" (UniqueName: \"kubernetes.io/projected/079821a6-579a-43df-bfbb-394c998cfac8-kube-api-access-f66fm\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246207 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246234 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246258 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246316 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246354 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246727 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.246821 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.247277 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.248079 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.248553 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.248599 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.248739 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.252867 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.256106 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-push\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.269312 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f66fm\" (UniqueName: \"kubernetes.io/projected/079821a6-579a-43df-bfbb-394c998cfac8-kube-api-access-f66fm\") pod \"service-telemetry-operator-5-build\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.419063 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:06 crc kubenswrapper[4750]: I0220 00:21:06.680559 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 20 00:21:07 crc kubenswrapper[4750]: I0220 00:21:07.119500 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"079821a6-579a-43df-bfbb-394c998cfac8","Type":"ContainerStarted","Data":"ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2"} Feb 20 00:21:07 crc kubenswrapper[4750]: I0220 00:21:07.119859 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"079821a6-579a-43df-bfbb-394c998cfac8","Type":"ContainerStarted","Data":"c6915acbe6d0bba3778b4471a4463bd0d49d5322c95cd85e9f48899ff95bc362"} Feb 20 00:21:07 crc kubenswrapper[4750]: E0220 00:21:07.174525 4750 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=2779325518433787891, SKID=, AKID=51:D2:4F:9F:30:C7:B6:3D:15:1B:65:4B:CC:7E:11:9A:B8:75:96:5A failed: x509: certificate signed by unknown authority" Feb 20 00:21:08 crc kubenswrapper[4750]: I0220 00:21:08.208451 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.135498 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-5-build" podUID="079821a6-579a-43df-bfbb-394c998cfac8" containerName="git-clone" containerID="cri-o://ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2" gracePeriod=30 Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.547386 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_079821a6-579a-43df-bfbb-394c998cfac8/git-clone/0.log" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.547688 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.592592 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-run\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.592662 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-buildcachedir\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.592716 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-ca-bundles\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.592769 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-buildworkdir\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.592870 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.592985 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.593591 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594077 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594238 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-push\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594330 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-root\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594403 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f66fm\" (UniqueName: \"kubernetes.io/projected/079821a6-579a-43df-bfbb-394c998cfac8-kube-api-access-f66fm\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594466 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-pull\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594513 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-node-pullsecrets\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594585 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-proxy-ca-bundles\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594642 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-system-configs\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.594676 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-build-blob-cache\") pod \"079821a6-579a-43df-bfbb-394c998cfac8\" (UID: \"079821a6-579a-43df-bfbb-394c998cfac8\") " Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.595334 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.595375 4750 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.595404 4750 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.595429 4750 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.595783 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.596808 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.595990 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.596628 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.596733 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.602615 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-pull" (OuterVolumeSpecName: "builder-dockercfg-gsktv-pull") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "builder-dockercfg-gsktv-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.602712 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/079821a6-579a-43df-bfbb-394c998cfac8-kube-api-access-f66fm" (OuterVolumeSpecName: "kube-api-access-f66fm") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "kube-api-access-f66fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.603350 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-push" (OuterVolumeSpecName: "builder-dockercfg-gsktv-push") pod "079821a6-579a-43df-bfbb-394c998cfac8" (UID: "079821a6-579a-43df-bfbb-394c998cfac8"). InnerVolumeSpecName "builder-dockercfg-gsktv-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696784 4750 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696813 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f66fm\" (UniqueName: \"kubernetes.io/projected/079821a6-579a-43df-bfbb-394c998cfac8-kube-api-access-f66fm\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696822 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-pull\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-pull\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696833 4750 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/079821a6-579a-43df-bfbb-394c998cfac8-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696842 4750 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696851 4750 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/079821a6-579a-43df-bfbb-394c998cfac8-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696859 4750 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/079821a6-579a-43df-bfbb-394c998cfac8-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:09 crc kubenswrapper[4750]: I0220 00:21:09.696867 4750 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-gsktv-push\" (UniqueName: \"kubernetes.io/secret/079821a6-579a-43df-bfbb-394c998cfac8-builder-dockercfg-gsktv-push\") on node \"crc\" DevicePath \"\"" Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.147051 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_079821a6-579a-43df-bfbb-394c998cfac8/git-clone/0.log" Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.147228 4750 generic.go:334] "Generic (PLEG): container finished" podID="079821a6-579a-43df-bfbb-394c998cfac8" containerID="ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2" exitCode=1 Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.147287 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"079821a6-579a-43df-bfbb-394c998cfac8","Type":"ContainerDied","Data":"ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2"} Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.147340 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"079821a6-579a-43df-bfbb-394c998cfac8","Type":"ContainerDied","Data":"c6915acbe6d0bba3778b4471a4463bd0d49d5322c95cd85e9f48899ff95bc362"} Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.147345 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.147379 4750 scope.go:117] "RemoveContainer" containerID="ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2" Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.178960 4750 scope.go:117] "RemoveContainer" containerID="ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2" Feb 20 00:21:10 crc kubenswrapper[4750]: E0220 00:21:10.179335 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2\": container with ID starting with ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2 not found: ID does not exist" containerID="ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2" Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.179371 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2"} err="failed to get container status \"ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2\": rpc error: code = NotFound desc = could not find container \"ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2\": container with ID starting with ff663c2fd8f7af33dc9355e0aadff5233bb373b5d5aea6db2093d62c3b93feb2 not found: ID does not exist" Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.194410 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.205422 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 20 00:21:10 crc kubenswrapper[4750]: I0220 00:21:10.574751 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="079821a6-579a-43df-bfbb-394c998cfac8" path="/var/lib/kubelet/pods/079821a6-579a-43df-bfbb-394c998cfac8/volumes" Feb 20 00:21:19 crc kubenswrapper[4750]: I0220 00:21:19.784930 4750 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 00:21:26 crc kubenswrapper[4750]: I0220 00:21:26.669758 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:21:26 crc kubenswrapper[4750]: I0220 00:21:26.670362 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.508224 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlzvm/must-gather-27zl2"] Feb 20 00:21:54 crc kubenswrapper[4750]: E0220 00:21:54.509211 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="079821a6-579a-43df-bfbb-394c998cfac8" containerName="git-clone" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.509233 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="079821a6-579a-43df-bfbb-394c998cfac8" containerName="git-clone" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.509399 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="079821a6-579a-43df-bfbb-394c998cfac8" containerName="git-clone" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.510434 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.515609 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zlzvm"/"openshift-service-ca.crt" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.515707 4750 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zlzvm"/"kube-root-ca.crt" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.526080 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zlzvm/must-gather-27zl2"] Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.586810 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snw4b\" (UniqueName: \"kubernetes.io/projected/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-kube-api-access-snw4b\") pod \"must-gather-27zl2\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.586900 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-must-gather-output\") pod \"must-gather-27zl2\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.687849 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-must-gather-output\") pod \"must-gather-27zl2\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.687960 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snw4b\" (UniqueName: \"kubernetes.io/projected/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-kube-api-access-snw4b\") pod \"must-gather-27zl2\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.688370 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-must-gather-output\") pod \"must-gather-27zl2\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.710892 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snw4b\" (UniqueName: \"kubernetes.io/projected/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-kube-api-access-snw4b\") pod \"must-gather-27zl2\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:54 crc kubenswrapper[4750]: I0220 00:21:54.826990 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:21:55 crc kubenswrapper[4750]: I0220 00:21:55.039062 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zlzvm/must-gather-27zl2"] Feb 20 00:21:55 crc kubenswrapper[4750]: I0220 00:21:55.505554 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlzvm/must-gather-27zl2" event={"ID":"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7","Type":"ContainerStarted","Data":"0028169d072d7b272402827498afd4bd0e25068c8d87afbf4a81c261f715e60e"} Feb 20 00:21:56 crc kubenswrapper[4750]: I0220 00:21:56.670424 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:21:56 crc kubenswrapper[4750]: I0220 00:21:56.670959 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:21:56 crc kubenswrapper[4750]: I0220 00:21:56.671011 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:21:56 crc kubenswrapper[4750]: I0220 00:21:56.672474 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2df5c0a1c7798ac15bd3ceb64e0a208a9488dd206f4f7114aee925599addd0eb"} pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 00:21:56 crc kubenswrapper[4750]: I0220 00:21:56.672572 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" containerID="cri-o://2df5c0a1c7798ac15bd3ceb64e0a208a9488dd206f4f7114aee925599addd0eb" gracePeriod=600 Feb 20 00:21:57 crc kubenswrapper[4750]: I0220 00:21:57.537401 4750 generic.go:334] "Generic (PLEG): container finished" podID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerID="2df5c0a1c7798ac15bd3ceb64e0a208a9488dd206f4f7114aee925599addd0eb" exitCode=0 Feb 20 00:21:57 crc kubenswrapper[4750]: I0220 00:21:57.537555 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerDied","Data":"2df5c0a1c7798ac15bd3ceb64e0a208a9488dd206f4f7114aee925599addd0eb"} Feb 20 00:21:57 crc kubenswrapper[4750]: I0220 00:21:57.537626 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"c56a2bad10e0c2524fa204d6534c1d8c4df69a01ec10d63bfc52aa010127a7d7"} Feb 20 00:21:57 crc kubenswrapper[4750]: I0220 00:21:57.537648 4750 scope.go:117] "RemoveContainer" containerID="4ba5fe5180b1882694da5fe6cd20529927f075cad71b329d6a7948bdcfb58cc2" Feb 20 00:22:03 crc kubenswrapper[4750]: I0220 00:22:03.580079 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlzvm/must-gather-27zl2" event={"ID":"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7","Type":"ContainerStarted","Data":"a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11"} Feb 20 00:22:03 crc kubenswrapper[4750]: I0220 00:22:03.580753 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlzvm/must-gather-27zl2" event={"ID":"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7","Type":"ContainerStarted","Data":"eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4"} Feb 20 00:22:03 crc kubenswrapper[4750]: I0220 00:22:03.607246 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zlzvm/must-gather-27zl2" podStartSLOduration=1.859738001 podStartE2EDuration="9.60721441s" podCreationTimestamp="2026-02-20 00:21:54 +0000 UTC" firstStartedPulling="2026-02-20 00:21:55.050448627 +0000 UTC m=+799.245284676" lastFinishedPulling="2026-02-20 00:22:02.797925036 +0000 UTC m=+806.992761085" observedRunningTime="2026-02-20 00:22:03.598080628 +0000 UTC m=+807.792916727" watchObservedRunningTime="2026-02-20 00:22:03.60721441 +0000 UTC m=+807.802050499" Feb 20 00:22:47 crc kubenswrapper[4750]: I0220 00:22:47.144174 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-z57xn_1c602074-9ed8-4385-a338-40fefebb4924/control-plane-machine-set-operator/0.log" Feb 20 00:22:47 crc kubenswrapper[4750]: I0220 00:22:47.183541 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gk5dc_347720b4-0171-4b11-9b37-940a17978ee1/kube-rbac-proxy/0.log" Feb 20 00:22:47 crc kubenswrapper[4750]: I0220 00:22:47.271826 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gk5dc_347720b4-0171-4b11-9b37-940a17978ee1/machine-api-operator/0.log" Feb 20 00:22:59 crc kubenswrapper[4750]: I0220 00:22:59.899097 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-pd5km_16f80a94-8f72-4baf-944f-8934f7a2cdc5/cert-manager-controller/0.log" Feb 20 00:23:00 crc kubenswrapper[4750]: I0220 00:23:00.099148 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-td2k7_34841f08-32f2-4738-8241-7b5c9e0408d7/cert-manager-webhook/0.log" Feb 20 00:23:00 crc kubenswrapper[4750]: I0220 00:23:00.110376 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-lg2cd_1eeff3f9-8995-450d-af18-d129c823ec09/cert-manager-cainjector/0.log" Feb 20 00:23:15 crc kubenswrapper[4750]: I0220 00:23:15.046676 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-6kn94_359c9e5a-f1ba-4513-9d7a-0903ae871923/prometheus-operator/0.log" Feb 20 00:23:15 crc kubenswrapper[4750]: I0220 00:23:15.232624 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89_df8960fb-1d2c-4857-988d-f51ea636c079/prometheus-operator-admission-webhook/0.log" Feb 20 00:23:15 crc kubenswrapper[4750]: I0220 00:23:15.253486 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r_3d27f900-e106-4a95-b60c-b75c81bf21e6/prometheus-operator-admission-webhook/0.log" Feb 20 00:23:15 crc kubenswrapper[4750]: I0220 00:23:15.447692 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-gsxjb_14f2560a-b7fb-491c-86ed-7b1e6fd77f5e/operator/0.log" Feb 20 00:23:15 crc kubenswrapper[4750]: I0220 00:23:15.516726 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-5vmsh_fc4519f1-96b2-4f48-a78c-10c8fda991a1/perses-operator/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.117244 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv_ebc81384-b7b6-4011-9040-d5655369377f/util/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.296985 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv_ebc81384-b7b6-4011-9040-d5655369377f/pull/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.297221 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv_ebc81384-b7b6-4011-9040-d5655369377f/pull/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.310561 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv_ebc81384-b7b6-4011-9040-d5655369377f/util/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.433880 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv_ebc81384-b7b6-4011-9040-d5655369377f/util/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.476927 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv_ebc81384-b7b6-4011-9040-d5655369377f/pull/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.503582 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1f7klv_ebc81384-b7b6-4011-9040-d5655369377f/extract/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.596767 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt_152dd755-10c1-482a-872d-31868f80c26f/util/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.762688 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt_152dd755-10c1-482a-872d-31868f80c26f/pull/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.767245 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt_152dd755-10c1-482a-872d-31868f80c26f/pull/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.780054 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt_152dd755-10c1-482a-872d-31868f80c26f/util/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.939655 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt_152dd755-10c1-482a-872d-31868f80c26f/pull/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.949989 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt_152dd755-10c1-482a-872d-31868f80c26f/util/0.log" Feb 20 00:23:31 crc kubenswrapper[4750]: I0220 00:23:31.985320 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdc5bt_152dd755-10c1-482a-872d-31868f80c26f/extract/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.124143 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl_3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf/util/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.293622 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl_3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf/pull/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.327587 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl_3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf/util/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.349188 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl_3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf/pull/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.529976 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl_3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf/pull/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.540842 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl_3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf/extract/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.573606 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5pq2bl_3b1fd4c3-b20e-48df-98fc-f0c08d5fdddf/util/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.696691 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c_744a8b27-0662-4e4f-8d91-1a2740182b2f/util/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.848510 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c_744a8b27-0662-4e4f-8d91-1a2740182b2f/util/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.871280 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c_744a8b27-0662-4e4f-8d91-1a2740182b2f/pull/0.log" Feb 20 00:23:32 crc kubenswrapper[4750]: I0220 00:23:32.906185 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c_744a8b27-0662-4e4f-8d91-1a2740182b2f/pull/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.011423 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c_744a8b27-0662-4e4f-8d91-1a2740182b2f/util/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.037184 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c_744a8b27-0662-4e4f-8d91-1a2740182b2f/pull/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.037406 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2f2c_744a8b27-0662-4e4f-8d91-1a2740182b2f/extract/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.187863 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5px4g_44fc4f50-400a-417e-9538-ef3dd5a66038/extract-utilities/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.330101 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5px4g_44fc4f50-400a-417e-9538-ef3dd5a66038/extract-utilities/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.338934 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5px4g_44fc4f50-400a-417e-9538-ef3dd5a66038/extract-content/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.375018 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5px4g_44fc4f50-400a-417e-9538-ef3dd5a66038/extract-content/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.518709 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5px4g_44fc4f50-400a-417e-9538-ef3dd5a66038/extract-utilities/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.518925 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5px4g_44fc4f50-400a-417e-9538-ef3dd5a66038/extract-content/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.694230 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5px4g_44fc4f50-400a-417e-9538-ef3dd5a66038/registry-server/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.750666 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mfjrj_2bfe5c0b-cde1-415e-846f-73947bfa62b4/extract-utilities/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.874501 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mfjrj_2bfe5c0b-cde1-415e-846f-73947bfa62b4/extract-utilities/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.898163 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mfjrj_2bfe5c0b-cde1-415e-846f-73947bfa62b4/extract-content/0.log" Feb 20 00:23:33 crc kubenswrapper[4750]: I0220 00:23:33.932957 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mfjrj_2bfe5c0b-cde1-415e-846f-73947bfa62b4/extract-content/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.063308 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mfjrj_2bfe5c0b-cde1-415e-846f-73947bfa62b4/extract-content/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.077328 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mfjrj_2bfe5c0b-cde1-415e-846f-73947bfa62b4/extract-utilities/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.274313 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-dvbwr_f26d2d81-da99-4a29-9f9f-f975eb3a75f3/marketplace-operator/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.320805 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mfjrj_2bfe5c0b-cde1-415e-846f-73947bfa62b4/registry-server/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.382228 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m6jfb_abac413a-7c83-4f73-8c5b-b1c0d50d5ebd/extract-utilities/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.518945 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m6jfb_abac413a-7c83-4f73-8c5b-b1c0d50d5ebd/extract-utilities/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.570465 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m6jfb_abac413a-7c83-4f73-8c5b-b1c0d50d5ebd/extract-content/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.592900 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m6jfb_abac413a-7c83-4f73-8c5b-b1c0d50d5ebd/extract-content/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.754639 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m6jfb_abac413a-7c83-4f73-8c5b-b1c0d50d5ebd/extract-utilities/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.757302 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m6jfb_abac413a-7c83-4f73-8c5b-b1c0d50d5ebd/extract-content/0.log" Feb 20 00:23:34 crc kubenswrapper[4750]: I0220 00:23:34.863834 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-m6jfb_abac413a-7c83-4f73-8c5b-b1c0d50d5ebd/registry-server/0.log" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.681314 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fb957"] Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.683967 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.714195 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fb957"] Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.814645 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w4mg\" (UniqueName: \"kubernetes.io/projected/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-kube-api-access-5w4mg\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.814724 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-catalog-content\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.814782 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-utilities\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.915743 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-catalog-content\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.915843 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-utilities\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.915864 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w4mg\" (UniqueName: \"kubernetes.io/projected/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-kube-api-access-5w4mg\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.916294 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-utilities\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.916795 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-catalog-content\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:41 crc kubenswrapper[4750]: I0220 00:23:41.934880 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w4mg\" (UniqueName: \"kubernetes.io/projected/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-kube-api-access-5w4mg\") pod \"community-operators-fb957\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:42 crc kubenswrapper[4750]: I0220 00:23:42.000361 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:42 crc kubenswrapper[4750]: I0220 00:23:42.536449 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fb957"] Feb 20 00:23:43 crc kubenswrapper[4750]: I0220 00:23:43.234901 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerID="4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b" exitCode=0 Feb 20 00:23:43 crc kubenswrapper[4750]: I0220 00:23:43.235001 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fb957" event={"ID":"4d6bc268-2a84-43ab-9cba-a369ab65d4bd","Type":"ContainerDied","Data":"4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b"} Feb 20 00:23:43 crc kubenswrapper[4750]: I0220 00:23:43.236952 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fb957" event={"ID":"4d6bc268-2a84-43ab-9cba-a369ab65d4bd","Type":"ContainerStarted","Data":"b0407ee2d67ac9720da595e94ebd1a53f058d65c46c0c8b826f34a23bef3713f"} Feb 20 00:23:44 crc kubenswrapper[4750]: I0220 00:23:44.244807 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fb957" event={"ID":"4d6bc268-2a84-43ab-9cba-a369ab65d4bd","Type":"ContainerStarted","Data":"4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5"} Feb 20 00:23:45 crc kubenswrapper[4750]: I0220 00:23:45.257804 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerID="4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5" exitCode=0 Feb 20 00:23:45 crc kubenswrapper[4750]: I0220 00:23:45.257853 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fb957" event={"ID":"4d6bc268-2a84-43ab-9cba-a369ab65d4bd","Type":"ContainerDied","Data":"4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5"} Feb 20 00:23:46 crc kubenswrapper[4750]: I0220 00:23:46.266058 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fb957" event={"ID":"4d6bc268-2a84-43ab-9cba-a369ab65d4bd","Type":"ContainerStarted","Data":"4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d"} Feb 20 00:23:46 crc kubenswrapper[4750]: I0220 00:23:46.286529 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fb957" podStartSLOduration=2.845551327 podStartE2EDuration="5.286509715s" podCreationTimestamp="2026-02-20 00:23:41 +0000 UTC" firstStartedPulling="2026-02-20 00:23:43.236579675 +0000 UTC m=+907.431415724" lastFinishedPulling="2026-02-20 00:23:45.677538023 +0000 UTC m=+909.872374112" observedRunningTime="2026-02-20 00:23:46.284715176 +0000 UTC m=+910.479551245" watchObservedRunningTime="2026-02-20 00:23:46.286509715 +0000 UTC m=+910.481345754" Feb 20 00:23:47 crc kubenswrapper[4750]: I0220 00:23:47.950589 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7cbcdd9f46-st58r_3d27f900-e106-4a95-b60c-b75c81bf21e6/prometheus-operator-admission-webhook/0.log" Feb 20 00:23:47 crc kubenswrapper[4750]: I0220 00:23:47.954036 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7cbcdd9f46-69f89_df8960fb-1d2c-4857-988d-f51ea636c079/prometheus-operator-admission-webhook/0.log" Feb 20 00:23:47 crc kubenswrapper[4750]: I0220 00:23:47.957764 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-6kn94_359c9e5a-f1ba-4513-9d7a-0903ae871923/prometheus-operator/0.log" Feb 20 00:23:48 crc kubenswrapper[4750]: I0220 00:23:48.043729 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-gsxjb_14f2560a-b7fb-491c-86ed-7b1e6fd77f5e/operator/0.log" Feb 20 00:23:48 crc kubenswrapper[4750]: I0220 00:23:48.114961 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-5vmsh_fc4519f1-96b2-4f48-a78c-10c8fda991a1/perses-operator/0.log" Feb 20 00:23:52 crc kubenswrapper[4750]: I0220 00:23:52.001072 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:52 crc kubenswrapper[4750]: I0220 00:23:52.002235 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:52 crc kubenswrapper[4750]: I0220 00:23:52.076008 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:52 crc kubenswrapper[4750]: I0220 00:23:52.353773 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:52 crc kubenswrapper[4750]: I0220 00:23:52.395442 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fb957"] Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.322222 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fb957" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="registry-server" containerID="cri-o://4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d" gracePeriod=2 Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.722718 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.744585 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xk2f7"] Feb 20 00:23:54 crc kubenswrapper[4750]: E0220 00:23:54.750266 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="registry-server" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.750289 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="registry-server" Feb 20 00:23:54 crc kubenswrapper[4750]: E0220 00:23:54.750303 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="extract-utilities" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.750309 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="extract-utilities" Feb 20 00:23:54 crc kubenswrapper[4750]: E0220 00:23:54.750325 4750 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="extract-content" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.750331 4750 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="extract-content" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.750431 4750 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerName="registry-server" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.751299 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.767267 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xk2f7"] Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.892150 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-utilities\") pod \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.892193 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-catalog-content\") pod \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.892271 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w4mg\" (UniqueName: \"kubernetes.io/projected/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-kube-api-access-5w4mg\") pod \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\" (UID: \"4d6bc268-2a84-43ab-9cba-a369ab65d4bd\") " Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.892502 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-utilities\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.892650 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-catalog-content\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.892864 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q57j6\" (UniqueName: \"kubernetes.io/projected/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-kube-api-access-q57j6\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.893196 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-utilities" (OuterVolumeSpecName: "utilities") pod "4d6bc268-2a84-43ab-9cba-a369ab65d4bd" (UID: "4d6bc268-2a84-43ab-9cba-a369ab65d4bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.903235 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-kube-api-access-5w4mg" (OuterVolumeSpecName: "kube-api-access-5w4mg") pod "4d6bc268-2a84-43ab-9cba-a369ab65d4bd" (UID: "4d6bc268-2a84-43ab-9cba-a369ab65d4bd"). InnerVolumeSpecName "kube-api-access-5w4mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.994513 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q57j6\" (UniqueName: \"kubernetes.io/projected/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-kube-api-access-q57j6\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.994588 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-utilities\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.994630 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-catalog-content\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.994758 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w4mg\" (UniqueName: \"kubernetes.io/projected/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-kube-api-access-5w4mg\") on node \"crc\" DevicePath \"\"" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.994775 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.995287 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-catalog-content\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:54 crc kubenswrapper[4750]: I0220 00:23:54.995344 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-utilities\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.014960 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q57j6\" (UniqueName: \"kubernetes.io/projected/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-kube-api-access-q57j6\") pod \"redhat-operators-xk2f7\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.072382 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.332550 4750 generic.go:334] "Generic (PLEG): container finished" podID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" containerID="4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d" exitCode=0 Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.332598 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fb957" event={"ID":"4d6bc268-2a84-43ab-9cba-a369ab65d4bd","Type":"ContainerDied","Data":"4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d"} Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.332626 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fb957" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.332647 4750 scope.go:117] "RemoveContainer" containerID="4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.332634 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fb957" event={"ID":"4d6bc268-2a84-43ab-9cba-a369ab65d4bd","Type":"ContainerDied","Data":"b0407ee2d67ac9720da595e94ebd1a53f058d65c46c0c8b826f34a23bef3713f"} Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.351709 4750 scope.go:117] "RemoveContainer" containerID="4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.374285 4750 scope.go:117] "RemoveContainer" containerID="4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.388894 4750 scope.go:117] "RemoveContainer" containerID="4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d" Feb 20 00:23:55 crc kubenswrapper[4750]: E0220 00:23:55.389464 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d\": container with ID starting with 4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d not found: ID does not exist" containerID="4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.389496 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d"} err="failed to get container status \"4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d\": rpc error: code = NotFound desc = could not find container \"4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d\": container with ID starting with 4aa29e567c80ad03d8bdef75adff8093c9ce9ab51c4dec75cdb0a54e3d5ff22d not found: ID does not exist" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.389516 4750 scope.go:117] "RemoveContainer" containerID="4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5" Feb 20 00:23:55 crc kubenswrapper[4750]: E0220 00:23:55.389816 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5\": container with ID starting with 4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5 not found: ID does not exist" containerID="4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.389840 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5"} err="failed to get container status \"4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5\": rpc error: code = NotFound desc = could not find container \"4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5\": container with ID starting with 4eab7c87e09b39e35042c81d8243166998c36c7edde49aa5207b74603624f0a5 not found: ID does not exist" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.389854 4750 scope.go:117] "RemoveContainer" containerID="4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b" Feb 20 00:23:55 crc kubenswrapper[4750]: E0220 00:23:55.390331 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b\": container with ID starting with 4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b not found: ID does not exist" containerID="4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.390348 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b"} err="failed to get container status \"4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b\": rpc error: code = NotFound desc = could not find container \"4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b\": container with ID starting with 4cfbb4df87835f0fca12ba376670bf16237b36c131f59cc568d272cd05ecb04b not found: ID does not exist" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.470898 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d6bc268-2a84-43ab-9cba-a369ab65d4bd" (UID: "4d6bc268-2a84-43ab-9cba-a369ab65d4bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.492630 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xk2f7"] Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.502248 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d6bc268-2a84-43ab-9cba-a369ab65d4bd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:23:55 crc kubenswrapper[4750]: W0220 00:23:55.505475 4750 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2a32fb5_6351_42fc_bfb2_8ec2f2f32e5c.slice/crio-412e20b992d7df137c805c02014c059aa85284ee7d674c8ba8a3540a902c7a15 WatchSource:0}: Error finding container 412e20b992d7df137c805c02014c059aa85284ee7d674c8ba8a3540a902c7a15: Status 404 returned error can't find the container with id 412e20b992d7df137c805c02014c059aa85284ee7d674c8ba8a3540a902c7a15 Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.665584 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fb957"] Feb 20 00:23:55 crc kubenswrapper[4750]: I0220 00:23:55.673871 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fb957"] Feb 20 00:23:56 crc kubenswrapper[4750]: I0220 00:23:56.341051 4750 generic.go:334] "Generic (PLEG): container finished" podID="f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" containerID="6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512" exitCode=0 Feb 20 00:23:56 crc kubenswrapper[4750]: I0220 00:23:56.341191 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk2f7" event={"ID":"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c","Type":"ContainerDied","Data":"6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512"} Feb 20 00:23:56 crc kubenswrapper[4750]: I0220 00:23:56.341278 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk2f7" event={"ID":"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c","Type":"ContainerStarted","Data":"412e20b992d7df137c805c02014c059aa85284ee7d674c8ba8a3540a902c7a15"} Feb 20 00:23:56 crc kubenswrapper[4750]: I0220 00:23:56.342638 4750 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 00:23:56 crc kubenswrapper[4750]: I0220 00:23:56.570911 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d6bc268-2a84-43ab-9cba-a369ab65d4bd" path="/var/lib/kubelet/pods/4d6bc268-2a84-43ab-9cba-a369ab65d4bd/volumes" Feb 20 00:23:56 crc kubenswrapper[4750]: I0220 00:23:56.670173 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:23:56 crc kubenswrapper[4750]: I0220 00:23:56.670232 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:23:57 crc kubenswrapper[4750]: I0220 00:23:57.348760 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk2f7" event={"ID":"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c","Type":"ContainerStarted","Data":"d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025"} Feb 20 00:23:58 crc kubenswrapper[4750]: I0220 00:23:58.357113 4750 generic.go:334] "Generic (PLEG): container finished" podID="f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" containerID="d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025" exitCode=0 Feb 20 00:23:58 crc kubenswrapper[4750]: I0220 00:23:58.357210 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk2f7" event={"ID":"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c","Type":"ContainerDied","Data":"d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025"} Feb 20 00:23:59 crc kubenswrapper[4750]: I0220 00:23:59.370567 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk2f7" event={"ID":"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c","Type":"ContainerStarted","Data":"0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d"} Feb 20 00:23:59 crc kubenswrapper[4750]: I0220 00:23:59.413206 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xk2f7" podStartSLOduration=2.7584630089999997 podStartE2EDuration="5.413074818s" podCreationTimestamp="2026-02-20 00:23:54 +0000 UTC" firstStartedPulling="2026-02-20 00:23:56.342431109 +0000 UTC m=+920.537267158" lastFinishedPulling="2026-02-20 00:23:58.997042888 +0000 UTC m=+923.191878967" observedRunningTime="2026-02-20 00:23:59.406287982 +0000 UTC m=+923.601124041" watchObservedRunningTime="2026-02-20 00:23:59.413074818 +0000 UTC m=+923.607910977" Feb 20 00:24:01 crc kubenswrapper[4750]: I0220 00:24:01.938943 4750 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jsxwq"] Feb 20 00:24:01 crc kubenswrapper[4750]: I0220 00:24:01.940286 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:01 crc kubenswrapper[4750]: I0220 00:24:01.957741 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jsxwq"] Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.098952 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-catalog-content\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.099003 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-utilities\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.099150 4750 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nvs5\" (UniqueName: \"kubernetes.io/projected/40ed8a4c-5643-4887-aaaf-107a325bb05a-kube-api-access-9nvs5\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.200348 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nvs5\" (UniqueName: \"kubernetes.io/projected/40ed8a4c-5643-4887-aaaf-107a325bb05a-kube-api-access-9nvs5\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.200451 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-catalog-content\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.200487 4750 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-utilities\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.201001 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-catalog-content\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.201310 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-utilities\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.225933 4750 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nvs5\" (UniqueName: \"kubernetes.io/projected/40ed8a4c-5643-4887-aaaf-107a325bb05a-kube-api-access-9nvs5\") pod \"certified-operators-jsxwq\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.260260 4750 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:02 crc kubenswrapper[4750]: I0220 00:24:02.499048 4750 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jsxwq"] Feb 20 00:24:03 crc kubenswrapper[4750]: I0220 00:24:03.415834 4750 generic.go:334] "Generic (PLEG): container finished" podID="40ed8a4c-5643-4887-aaaf-107a325bb05a" containerID="b8d58c0a3318393ac88ee322f02feea4ffa639a30deea5d40b1409b165a256a6" exitCode=0 Feb 20 00:24:03 crc kubenswrapper[4750]: I0220 00:24:03.416141 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jsxwq" event={"ID":"40ed8a4c-5643-4887-aaaf-107a325bb05a","Type":"ContainerDied","Data":"b8d58c0a3318393ac88ee322f02feea4ffa639a30deea5d40b1409b165a256a6"} Feb 20 00:24:03 crc kubenswrapper[4750]: I0220 00:24:03.416170 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jsxwq" event={"ID":"40ed8a4c-5643-4887-aaaf-107a325bb05a","Type":"ContainerStarted","Data":"f3a7bf6f7054dcf23b3fa84c5241ee960734cadf642096ad4bd438a597eb0bdc"} Feb 20 00:24:05 crc kubenswrapper[4750]: I0220 00:24:05.073172 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:24:05 crc kubenswrapper[4750]: I0220 00:24:05.074085 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:24:05 crc kubenswrapper[4750]: I0220 00:24:05.434297 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jsxwq" event={"ID":"40ed8a4c-5643-4887-aaaf-107a325bb05a","Type":"ContainerStarted","Data":"75c4836c92cef28f0062c725e0b5621d10c39d2ffb92ab53329787c02f25b4f6"} Feb 20 00:24:06 crc kubenswrapper[4750]: I0220 00:24:06.122465 4750 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xk2f7" podUID="f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" containerName="registry-server" probeResult="failure" output=< Feb 20 00:24:06 crc kubenswrapper[4750]: timeout: failed to connect service ":50051" within 1s Feb 20 00:24:06 crc kubenswrapper[4750]: > Feb 20 00:24:06 crc kubenswrapper[4750]: I0220 00:24:06.440291 4750 generic.go:334] "Generic (PLEG): container finished" podID="40ed8a4c-5643-4887-aaaf-107a325bb05a" containerID="75c4836c92cef28f0062c725e0b5621d10c39d2ffb92ab53329787c02f25b4f6" exitCode=0 Feb 20 00:24:06 crc kubenswrapper[4750]: I0220 00:24:06.440359 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jsxwq" event={"ID":"40ed8a4c-5643-4887-aaaf-107a325bb05a","Type":"ContainerDied","Data":"75c4836c92cef28f0062c725e0b5621d10c39d2ffb92ab53329787c02f25b4f6"} Feb 20 00:24:07 crc kubenswrapper[4750]: I0220 00:24:07.448593 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jsxwq" event={"ID":"40ed8a4c-5643-4887-aaaf-107a325bb05a","Type":"ContainerStarted","Data":"630aec47766fa15860ce4aa2f0130160409a24d29b513bb9282126f3666a0a5c"} Feb 20 00:24:07 crc kubenswrapper[4750]: I0220 00:24:07.464043 4750 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jsxwq" podStartSLOduration=2.816593484 podStartE2EDuration="6.464025432s" podCreationTimestamp="2026-02-20 00:24:01 +0000 UTC" firstStartedPulling="2026-02-20 00:24:03.417455857 +0000 UTC m=+927.612291906" lastFinishedPulling="2026-02-20 00:24:07.064887795 +0000 UTC m=+931.259723854" observedRunningTime="2026-02-20 00:24:07.462410478 +0000 UTC m=+931.657246547" watchObservedRunningTime="2026-02-20 00:24:07.464025432 +0000 UTC m=+931.658861481" Feb 20 00:24:12 crc kubenswrapper[4750]: I0220 00:24:12.261927 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:12 crc kubenswrapper[4750]: I0220 00:24:12.262634 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:12 crc kubenswrapper[4750]: I0220 00:24:12.347244 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:12 crc kubenswrapper[4750]: I0220 00:24:12.537000 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:12 crc kubenswrapper[4750]: I0220 00:24:12.883594 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jsxwq"] Feb 20 00:24:14 crc kubenswrapper[4750]: I0220 00:24:14.497042 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jsxwq" podUID="40ed8a4c-5643-4887-aaaf-107a325bb05a" containerName="registry-server" containerID="cri-o://630aec47766fa15860ce4aa2f0130160409a24d29b513bb9282126f3666a0a5c" gracePeriod=2 Feb 20 00:24:15 crc kubenswrapper[4750]: I0220 00:24:15.143488 4750 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:24:15 crc kubenswrapper[4750]: I0220 00:24:15.207000 4750 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.282037 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xk2f7"] Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.514469 4750 generic.go:334] "Generic (PLEG): container finished" podID="40ed8a4c-5643-4887-aaaf-107a325bb05a" containerID="630aec47766fa15860ce4aa2f0130160409a24d29b513bb9282126f3666a0a5c" exitCode=0 Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.514541 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jsxwq" event={"ID":"40ed8a4c-5643-4887-aaaf-107a325bb05a","Type":"ContainerDied","Data":"630aec47766fa15860ce4aa2f0130160409a24d29b513bb9282126f3666a0a5c"} Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.514851 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xk2f7" podUID="f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" containerName="registry-server" containerID="cri-o://0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d" gracePeriod=2 Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.834858 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.973012 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-utilities\") pod \"40ed8a4c-5643-4887-aaaf-107a325bb05a\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.973409 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-catalog-content\") pod \"40ed8a4c-5643-4887-aaaf-107a325bb05a\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.973507 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nvs5\" (UniqueName: \"kubernetes.io/projected/40ed8a4c-5643-4887-aaaf-107a325bb05a-kube-api-access-9nvs5\") pod \"40ed8a4c-5643-4887-aaaf-107a325bb05a\" (UID: \"40ed8a4c-5643-4887-aaaf-107a325bb05a\") " Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.973901 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-utilities" (OuterVolumeSpecName: "utilities") pod "40ed8a4c-5643-4887-aaaf-107a325bb05a" (UID: "40ed8a4c-5643-4887-aaaf-107a325bb05a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.975840 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:16 crc kubenswrapper[4750]: I0220 00:24:16.991508 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ed8a4c-5643-4887-aaaf-107a325bb05a-kube-api-access-9nvs5" (OuterVolumeSpecName: "kube-api-access-9nvs5") pod "40ed8a4c-5643-4887-aaaf-107a325bb05a" (UID: "40ed8a4c-5643-4887-aaaf-107a325bb05a"). InnerVolumeSpecName "kube-api-access-9nvs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.058032 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40ed8a4c-5643-4887-aaaf-107a325bb05a" (UID: "40ed8a4c-5643-4887-aaaf-107a325bb05a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.077718 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ed8a4c-5643-4887-aaaf-107a325bb05a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.077802 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nvs5\" (UniqueName: \"kubernetes.io/projected/40ed8a4c-5643-4887-aaaf-107a325bb05a-kube-api-access-9nvs5\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.469498 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.522278 4750 generic.go:334] "Generic (PLEG): container finished" podID="f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" containerID="0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d" exitCode=0 Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.522341 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk2f7" event={"ID":"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c","Type":"ContainerDied","Data":"0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d"} Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.522366 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xk2f7" event={"ID":"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c","Type":"ContainerDied","Data":"412e20b992d7df137c805c02014c059aa85284ee7d674c8ba8a3540a902c7a15"} Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.522382 4750 scope.go:117] "RemoveContainer" containerID="0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.522647 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xk2f7" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.528403 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jsxwq" event={"ID":"40ed8a4c-5643-4887-aaaf-107a325bb05a","Type":"ContainerDied","Data":"f3a7bf6f7054dcf23b3fa84c5241ee960734cadf642096ad4bd438a597eb0bdc"} Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.528616 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jsxwq" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.546032 4750 scope.go:117] "RemoveContainer" containerID="d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.575285 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jsxwq"] Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.579425 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jsxwq"] Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.591816 4750 scope.go:117] "RemoveContainer" containerID="6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.592881 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-catalog-content\") pod \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.592987 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-utilities\") pod \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.593075 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q57j6\" (UniqueName: \"kubernetes.io/projected/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-kube-api-access-q57j6\") pod \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\" (UID: \"f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c\") " Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.594802 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-utilities" (OuterVolumeSpecName: "utilities") pod "f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" (UID: "f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.597806 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-kube-api-access-q57j6" (OuterVolumeSpecName: "kube-api-access-q57j6") pod "f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" (UID: "f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c"). InnerVolumeSpecName "kube-api-access-q57j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.618373 4750 scope.go:117] "RemoveContainer" containerID="0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d" Feb 20 00:24:17 crc kubenswrapper[4750]: E0220 00:24:17.618778 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d\": container with ID starting with 0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d not found: ID does not exist" containerID="0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.618804 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d"} err="failed to get container status \"0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d\": rpc error: code = NotFound desc = could not find container \"0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d\": container with ID starting with 0e87f724e0abd581dcfc6c99de27a5f37c04c9f47d45254adb1076a7dc876a5d not found: ID does not exist" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.618822 4750 scope.go:117] "RemoveContainer" containerID="d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025" Feb 20 00:24:17 crc kubenswrapper[4750]: E0220 00:24:17.619026 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025\": container with ID starting with d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025 not found: ID does not exist" containerID="d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.619048 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025"} err="failed to get container status \"d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025\": rpc error: code = NotFound desc = could not find container \"d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025\": container with ID starting with d682a940cfe5f1a4921d98cfe8654e4f34ec64330b71f28e6ffcfa85bee3d025 not found: ID does not exist" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.619060 4750 scope.go:117] "RemoveContainer" containerID="6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512" Feb 20 00:24:17 crc kubenswrapper[4750]: E0220 00:24:17.619308 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512\": container with ID starting with 6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512 not found: ID does not exist" containerID="6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.619333 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512"} err="failed to get container status \"6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512\": rpc error: code = NotFound desc = could not find container \"6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512\": container with ID starting with 6f1a18460219a065e0bea85cde460215e05a119f364b6794fd047144be55f512 not found: ID does not exist" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.619348 4750 scope.go:117] "RemoveContainer" containerID="630aec47766fa15860ce4aa2f0130160409a24d29b513bb9282126f3666a0a5c" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.647390 4750 scope.go:117] "RemoveContainer" containerID="75c4836c92cef28f0062c725e0b5621d10c39d2ffb92ab53329787c02f25b4f6" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.672224 4750 scope.go:117] "RemoveContainer" containerID="b8d58c0a3318393ac88ee322f02feea4ffa639a30deea5d40b1409b165a256a6" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.694953 4750 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.694978 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q57j6\" (UniqueName: \"kubernetes.io/projected/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-kube-api-access-q57j6\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.767286 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" (UID: "f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.796770 4750 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.858850 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xk2f7"] Feb 20 00:24:17 crc kubenswrapper[4750]: I0220 00:24:17.863575 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xk2f7"] Feb 20 00:24:18 crc kubenswrapper[4750]: I0220 00:24:18.568378 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40ed8a4c-5643-4887-aaaf-107a325bb05a" path="/var/lib/kubelet/pods/40ed8a4c-5643-4887-aaaf-107a325bb05a/volumes" Feb 20 00:24:18 crc kubenswrapper[4750]: I0220 00:24:18.569799 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c" path="/var/lib/kubelet/pods/f2a32fb5-6351-42fc-bfb2-8ec2f2f32e5c/volumes" Feb 20 00:24:26 crc kubenswrapper[4750]: I0220 00:24:26.669364 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:24:26 crc kubenswrapper[4750]: I0220 00:24:26.669947 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:24:37 crc kubenswrapper[4750]: I0220 00:24:37.726902 4750 generic.go:334] "Generic (PLEG): container finished" podID="dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7" containerID="eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4" exitCode=0 Feb 20 00:24:37 crc kubenswrapper[4750]: I0220 00:24:37.726968 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlzvm/must-gather-27zl2" event={"ID":"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7","Type":"ContainerDied","Data":"eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4"} Feb 20 00:24:37 crc kubenswrapper[4750]: I0220 00:24:37.728179 4750 scope.go:117] "RemoveContainer" containerID="eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4" Feb 20 00:24:38 crc kubenswrapper[4750]: I0220 00:24:38.068348 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlzvm_must-gather-27zl2_dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7/gather/0.log" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.108410 4750 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlzvm/must-gather-27zl2"] Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.109428 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zlzvm/must-gather-27zl2" podUID="dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7" containerName="copy" containerID="cri-o://a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11" gracePeriod=2 Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.114905 4750 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlzvm/must-gather-27zl2"] Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.452922 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlzvm_must-gather-27zl2_dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7/copy/0.log" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.453817 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.609606 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-must-gather-output\") pod \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.609665 4750 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snw4b\" (UniqueName: \"kubernetes.io/projected/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-kube-api-access-snw4b\") pod \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\" (UID: \"dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7\") " Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.615931 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-kube-api-access-snw4b" (OuterVolumeSpecName: "kube-api-access-snw4b") pod "dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7" (UID: "dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7"). InnerVolumeSpecName "kube-api-access-snw4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.661526 4750 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7" (UID: "dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.712216 4750 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.712272 4750 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snw4b\" (UniqueName: \"kubernetes.io/projected/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7-kube-api-access-snw4b\") on node \"crc\" DevicePath \"\"" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.791799 4750 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlzvm_must-gather-27zl2_dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7/copy/0.log" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.792370 4750 generic.go:334] "Generic (PLEG): container finished" podID="dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7" containerID="a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11" exitCode=143 Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.792464 4750 scope.go:117] "RemoveContainer" containerID="a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.792494 4750 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlzvm/must-gather-27zl2" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.820507 4750 scope.go:117] "RemoveContainer" containerID="eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.875935 4750 scope.go:117] "RemoveContainer" containerID="a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11" Feb 20 00:24:45 crc kubenswrapper[4750]: E0220 00:24:45.876419 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11\": container with ID starting with a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11 not found: ID does not exist" containerID="a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.876450 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11"} err="failed to get container status \"a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11\": rpc error: code = NotFound desc = could not find container \"a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11\": container with ID starting with a246c1bfc49b917307c9d9fd47c1f807e4eb985ceae885d92141cd328e529c11 not found: ID does not exist" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.876473 4750 scope.go:117] "RemoveContainer" containerID="eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4" Feb 20 00:24:45 crc kubenswrapper[4750]: E0220 00:24:45.876768 4750 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4\": container with ID starting with eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4 not found: ID does not exist" containerID="eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4" Feb 20 00:24:45 crc kubenswrapper[4750]: I0220 00:24:45.876834 4750 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4"} err="failed to get container status \"eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4\": rpc error: code = NotFound desc = could not find container \"eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4\": container with ID starting with eb64753151cf824d4ab39255ebb3dcbc3e96ba2017c9cc46e4bade07b091e5f4 not found: ID does not exist" Feb 20 00:24:46 crc kubenswrapper[4750]: I0220 00:24:46.568785 4750 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7" path="/var/lib/kubelet/pods/dc77ebe4-bfe4-4ba6-9318-b3c81d6856e7/volumes" Feb 20 00:24:56 crc kubenswrapper[4750]: I0220 00:24:56.670161 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:24:56 crc kubenswrapper[4750]: I0220 00:24:56.670679 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:24:56 crc kubenswrapper[4750]: I0220 00:24:56.670730 4750 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" Feb 20 00:24:56 crc kubenswrapper[4750]: I0220 00:24:56.671324 4750 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c56a2bad10e0c2524fa204d6534c1d8c4df69a01ec10d63bfc52aa010127a7d7"} pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 00:24:56 crc kubenswrapper[4750]: I0220 00:24:56.671386 4750 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" containerID="cri-o://c56a2bad10e0c2524fa204d6534c1d8c4df69a01ec10d63bfc52aa010127a7d7" gracePeriod=600 Feb 20 00:24:57 crc kubenswrapper[4750]: I0220 00:24:57.879313 4750 generic.go:334] "Generic (PLEG): container finished" podID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerID="c56a2bad10e0c2524fa204d6534c1d8c4df69a01ec10d63bfc52aa010127a7d7" exitCode=0 Feb 20 00:24:57 crc kubenswrapper[4750]: I0220 00:24:57.879874 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerDied","Data":"c56a2bad10e0c2524fa204d6534c1d8c4df69a01ec10d63bfc52aa010127a7d7"} Feb 20 00:24:57 crc kubenswrapper[4750]: I0220 00:24:57.879905 4750 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" event={"ID":"62d0755e-ca22-4187-aae7-65377cedcaa9","Type":"ContainerStarted","Data":"03d0833a6799ece9d80f155f4322b3ad76c80edadd4417d7b594e0efa67074bb"} Feb 20 00:24:57 crc kubenswrapper[4750]: I0220 00:24:57.879925 4750 scope.go:117] "RemoveContainer" containerID="2df5c0a1c7798ac15bd3ceb64e0a208a9488dd206f4f7114aee925599addd0eb" Feb 20 00:27:26 crc kubenswrapper[4750]: I0220 00:27:26.670628 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:27:26 crc kubenswrapper[4750]: I0220 00:27:26.671497 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 00:27:56 crc kubenswrapper[4750]: I0220 00:27:56.669461 4750 patch_prober.go:28] interesting pod/machine-config-daemon-fk2zg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 00:27:56 crc kubenswrapper[4750]: I0220 00:27:56.670088 4750 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fk2zg" podUID="62d0755e-ca22-4187-aae7-65377cedcaa9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515145725240024452 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015145725241017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015145722206016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015145722206015461 5ustar corecore